var/home/core/zuul-output/0000755000175000017500000000000015111421770014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111426203015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003767475215111426175017725 0ustar rootrootNov 25 21:51:00 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 21:51:00 crc restorecon[4682]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:00 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 21:51:01 crc restorecon[4682]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 21:51:02 crc kubenswrapper[4931]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 21:51:02 crc kubenswrapper[4931]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 21:51:02 crc kubenswrapper[4931]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 21:51:02 crc kubenswrapper[4931]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 21:51:02 crc kubenswrapper[4931]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 21:51:02 crc kubenswrapper[4931]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.123023 4931 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129301 4931 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129331 4931 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129341 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129350 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129357 4931 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129366 4931 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129374 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129383 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129392 4931 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129400 4931 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129408 4931 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129416 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129424 4931 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129432 4931 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129439 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129447 4931 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129455 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129463 4931 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129470 4931 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129485 4931 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129493 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129501 4931 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129509 4931 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129516 4931 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129524 4931 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129531 4931 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129539 4931 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129547 4931 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129554 4931 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129562 4931 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129571 4931 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129582 4931 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129591 4931 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129601 4931 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129615 4931 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129628 4931 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129642 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129654 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129665 4931 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129675 4931 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129688 4931 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129700 4931 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129709 4931 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129717 4931 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129728 4931 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129737 4931 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129746 4931 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129755 4931 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129763 4931 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129770 4931 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129778 4931 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129785 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129793 4931 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129802 4931 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129811 4931 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129821 4931 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129831 4931 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129840 4931 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129848 4931 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129856 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129864 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129904 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129915 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129924 4931 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129932 4931 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129940 4931 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129948 4931 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129956 4931 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129964 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129972 4931 feature_gate.go:330] unrecognized feature gate: Example Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.129980 4931 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131044 4931 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131083 4931 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131105 4931 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131120 4931 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131134 4931 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131145 4931 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131160 4931 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131173 4931 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131186 4931 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131198 4931 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131210 4931 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131222 4931 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131234 4931 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131246 4931 flags.go:64] FLAG: --cgroup-root="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131258 4931 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131269 4931 flags.go:64] FLAG: --client-ca-file="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131281 4931 flags.go:64] FLAG: --cloud-config="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131292 4931 flags.go:64] FLAG: --cloud-provider="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131303 4931 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131317 4931 flags.go:64] FLAG: --cluster-domain="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131329 4931 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131341 4931 flags.go:64] FLAG: --config-dir="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131352 4931 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131365 4931 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131381 4931 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131393 4931 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131405 4931 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131417 4931 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131428 4931 flags.go:64] FLAG: --contention-profiling="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131440 4931 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131452 4931 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131464 4931 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131475 4931 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131490 4931 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131502 4931 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131513 4931 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131524 4931 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131538 4931 flags.go:64] FLAG: --enable-server="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131549 4931 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131566 4931 flags.go:64] FLAG: --event-burst="100" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131577 4931 flags.go:64] FLAG: --event-qps="50" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131586 4931 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131595 4931 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131604 4931 flags.go:64] FLAG: --eviction-hard="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131616 4931 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131624 4931 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131633 4931 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131644 4931 flags.go:64] FLAG: --eviction-soft="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131654 4931 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131665 4931 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131677 4931 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131688 4931 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131699 4931 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131710 4931 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131721 4931 flags.go:64] FLAG: --feature-gates="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131736 4931 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131748 4931 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131759 4931 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131772 4931 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131783 4931 flags.go:64] FLAG: --healthz-port="10248" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131795 4931 flags.go:64] FLAG: --help="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131807 4931 flags.go:64] FLAG: --hostname-override="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131821 4931 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131834 4931 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131846 4931 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131858 4931 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131869 4931 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131928 4931 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131941 4931 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131952 4931 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131963 4931 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131975 4931 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.131987 4931 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132000 4931 flags.go:64] FLAG: --kube-reserved="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132012 4931 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132023 4931 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132035 4931 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132045 4931 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132054 4931 flags.go:64] FLAG: --lock-file="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132063 4931 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132072 4931 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132082 4931 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132096 4931 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132105 4931 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132114 4931 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132123 4931 flags.go:64] FLAG: --logging-format="text" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132132 4931 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132142 4931 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132150 4931 flags.go:64] FLAG: --manifest-url="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132159 4931 flags.go:64] FLAG: --manifest-url-header="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132170 4931 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132180 4931 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132190 4931 flags.go:64] FLAG: --max-pods="110" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132199 4931 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132208 4931 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132216 4931 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132227 4931 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132275 4931 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132285 4931 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132294 4931 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132315 4931 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132324 4931 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132335 4931 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132344 4931 flags.go:64] FLAG: --pod-cidr="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132353 4931 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132367 4931 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132377 4931 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132386 4931 flags.go:64] FLAG: --pods-per-core="0" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132395 4931 flags.go:64] FLAG: --port="10250" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132406 4931 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132415 4931 flags.go:64] FLAG: --provider-id="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132424 4931 flags.go:64] FLAG: --qos-reserved="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132433 4931 flags.go:64] FLAG: --read-only-port="10255" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132443 4931 flags.go:64] FLAG: --register-node="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132451 4931 flags.go:64] FLAG: --register-schedulable="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132460 4931 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132474 4931 flags.go:64] FLAG: --registry-burst="10" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132483 4931 flags.go:64] FLAG: --registry-qps="5" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132492 4931 flags.go:64] FLAG: --reserved-cpus="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132501 4931 flags.go:64] FLAG: --reserved-memory="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132512 4931 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132522 4931 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132531 4931 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132540 4931 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132549 4931 flags.go:64] FLAG: --runonce="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132557 4931 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132566 4931 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132576 4931 flags.go:64] FLAG: --seccomp-default="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132585 4931 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132594 4931 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132603 4931 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132612 4931 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132621 4931 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132630 4931 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132639 4931 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132648 4931 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132657 4931 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132666 4931 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132675 4931 flags.go:64] FLAG: --system-cgroups="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132684 4931 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132697 4931 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132706 4931 flags.go:64] FLAG: --tls-cert-file="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132715 4931 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132726 4931 flags.go:64] FLAG: --tls-min-version="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132734 4931 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132744 4931 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132754 4931 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132766 4931 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132778 4931 flags.go:64] FLAG: --v="2" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132792 4931 flags.go:64] FLAG: --version="false" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132817 4931 flags.go:64] FLAG: --vmodule="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132831 4931 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.132844 4931 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133095 4931 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133108 4931 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133118 4931 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133127 4931 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133135 4931 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133143 4931 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133151 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133160 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133168 4931 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133178 4931 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133188 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133197 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133205 4931 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133214 4931 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133222 4931 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133230 4931 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133238 4931 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133246 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133253 4931 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133264 4931 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133273 4931 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133281 4931 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133291 4931 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133299 4931 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133307 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133319 4931 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133328 4931 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133337 4931 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133347 4931 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133355 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133366 4931 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133376 4931 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133385 4931 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133393 4931 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133401 4931 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133410 4931 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133418 4931 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133426 4931 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133434 4931 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133441 4931 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133449 4931 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133457 4931 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133472 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133480 4931 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133488 4931 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133496 4931 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133503 4931 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133511 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133519 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133527 4931 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133534 4931 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133542 4931 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133550 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133559 4931 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133566 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133574 4931 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133582 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133591 4931 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133600 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133607 4931 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133615 4931 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133623 4931 feature_gate.go:330] unrecognized feature gate: Example Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133631 4931 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133639 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133648 4931 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133656 4931 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133663 4931 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133671 4931 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133679 4931 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133687 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.133694 4931 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.133723 4931 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.147675 4931 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.147733 4931 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.147931 4931 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.147969 4931 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.147988 4931 feature_gate.go:330] unrecognized feature gate: Example Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148001 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148012 4931 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148025 4931 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148034 4931 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148043 4931 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148051 4931 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148059 4931 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148067 4931 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148075 4931 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148083 4931 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148091 4931 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148099 4931 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148107 4931 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148116 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148124 4931 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148133 4931 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148141 4931 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148149 4931 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148157 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148165 4931 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148173 4931 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148180 4931 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148189 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148196 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148204 4931 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148213 4931 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148220 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148228 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148241 4931 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148251 4931 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148261 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148271 4931 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148280 4931 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148288 4931 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148298 4931 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148308 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148316 4931 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148359 4931 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148367 4931 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148375 4931 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148383 4931 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148391 4931 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148399 4931 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148406 4931 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148415 4931 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148422 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148431 4931 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148441 4931 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148451 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148461 4931 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148469 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148478 4931 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148486 4931 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148494 4931 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148502 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148510 4931 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148518 4931 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148526 4931 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148533 4931 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148541 4931 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148550 4931 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148557 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148566 4931 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148573 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148581 4931 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148591 4931 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148600 4931 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148610 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.148623 4931 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148874 4931 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148926 4931 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148938 4931 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148950 4931 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148960 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148970 4931 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148979 4931 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148987 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.148996 4931 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149006 4931 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149016 4931 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149024 4931 feature_gate.go:330] unrecognized feature gate: Example Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149033 4931 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149042 4931 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149051 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149059 4931 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149068 4931 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149075 4931 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149084 4931 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149091 4931 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149099 4931 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149106 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149114 4931 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149122 4931 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149129 4931 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149138 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149146 4931 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149155 4931 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149164 4931 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149172 4931 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149181 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149189 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149197 4931 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149205 4931 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149214 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149222 4931 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149229 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149237 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149245 4931 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149253 4931 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149263 4931 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149273 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149281 4931 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149290 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149299 4931 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149307 4931 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149315 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149323 4931 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149331 4931 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149339 4931 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149347 4931 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149355 4931 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149363 4931 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149370 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149378 4931 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149386 4931 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149393 4931 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149401 4931 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149411 4931 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149421 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149429 4931 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149438 4931 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149446 4931 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149455 4931 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149463 4931 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149472 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149479 4931 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149487 4931 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149495 4931 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149502 4931 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.149511 4931 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.149522 4931 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.149746 4931 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.155432 4931 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.155553 4931 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.157537 4931 server.go:997] "Starting client certificate rotation" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.157590 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.157826 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-17 12:28:51.638692982 +0000 UTC Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.158015 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.187968 4931 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.191946 4931 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.192186 4931 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.214187 4931 log.go:25] "Validated CRI v1 runtime API" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.268711 4931 log.go:25] "Validated CRI v1 image API" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.271566 4931 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.277120 4931 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-21-46-24-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.277182 4931 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.308598 4931 manager.go:217] Machine: {Timestamp:2025-11-25 21:51:02.306092628 +0000 UTC m=+0.662604827 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:dfb0006d-ded5-40c5-a63e-512b52836fc8 BootID:80babd60-bb40-45e4-b0a0-67856b72145a Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:94:a9:cf Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:94:a9:cf Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:93:e7:44 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:d8:32:ff Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:84:12:ca Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d3:7f:ef Speed:-1 Mtu:1496} {Name:eth10 MacAddress:72:58:d3:df:8c:ad Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ce:d3:a0:63:58:49 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.309023 4931 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.309261 4931 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.311338 4931 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.311682 4931 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.311734 4931 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.312128 4931 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.312146 4931 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.312722 4931 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.312775 4931 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.313207 4931 state_mem.go:36] "Initialized new in-memory state store" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.313359 4931 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.319741 4931 kubelet.go:418] "Attempting to sync node with API server" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.319778 4931 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.319820 4931 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.319842 4931 kubelet.go:324] "Adding apiserver pod source" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.319862 4931 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.324699 4931 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.325794 4931 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.326622 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.326674 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.326832 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.326757 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.328563 4931 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330467 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330529 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330551 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330571 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330600 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330639 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330657 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330687 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330723 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330743 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330796 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330829 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.330934 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.331707 4931 server.go:1280] "Started kubelet" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.332480 4931 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.333378 4931 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.333633 4931 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 21:51:02 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.337737 4931 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.342010 4931 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.342123 4931 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.342142 4931 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 02:39:12.091416969 +0000 UTC Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.342386 4931 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.342400 4931 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.342679 4931 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.342724 4931 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.341848 4931 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b5e63ede4b69f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 21:51:02.331643551 +0000 UTC m=+0.688155710,LastTimestamp:2025-11-25 21:51:02.331643551 +0000 UTC m=+0.688155710,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.343871 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.344009 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.344733 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="200ms" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.344773 4931 server.go:460] "Adding debug handlers to kubelet server" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.346769 4931 factory.go:55] Registering systemd factory Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.346811 4931 factory.go:221] Registration of the systemd container factory successfully Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.347754 4931 factory.go:153] Registering CRI-O factory Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.347804 4931 factory.go:221] Registration of the crio container factory successfully Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.347982 4931 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.348038 4931 factory.go:103] Registering Raw factory Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.348081 4931 manager.go:1196] Started watching for new ooms in manager Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.349458 4931 manager.go:319] Starting recovery of all containers Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365516 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365607 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365636 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365657 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365683 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365704 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365723 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365743 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365767 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365788 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365810 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365840 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365858 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365950 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365974 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.365993 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366012 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366031 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366050 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366070 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366089 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366110 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366130 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366149 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366216 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366237 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366264 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366288 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366310 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366331 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366350 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366368 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366390 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366409 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366464 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366487 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366506 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366525 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366545 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366563 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366581 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366634 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366654 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366673 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366692 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366713 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366734 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366752 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366772 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366789 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366808 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366826 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366854 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366923 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366958 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.366981 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367004 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367024 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367043 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367062 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367084 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367105 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367124 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367143 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367162 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367251 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367272 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367291 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367311 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367330 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367348 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367369 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367389 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367435 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367457 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367476 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367495 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367515 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367532 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367550 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367566 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367584 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367603 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367622 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367643 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367662 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367680 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367699 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367717 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367738 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367756 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367776 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367795 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367816 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367834 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367856 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367913 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.367948 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368011 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368030 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368050 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368069 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368089 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368109 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368138 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368158 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368178 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368198 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368218 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368239 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368259 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368281 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368302 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368320 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368340 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368373 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368404 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368425 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368443 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368461 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368483 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368500 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368517 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368536 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368555 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368575 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368596 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368613 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368634 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368658 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368676 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368727 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368745 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368764 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368784 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368802 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368821 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368838 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368858 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368946 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368969 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.368992 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369014 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369033 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369051 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369069 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369088 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369108 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369128 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369146 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369165 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369185 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369202 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369221 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369252 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369269 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369286 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369305 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369324 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369345 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369364 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369392 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369412 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369433 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.369453 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.374804 4931 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.374869 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.374920 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.374943 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.374962 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.374987 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375003 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375021 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375042 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375057 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375075 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375100 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375117 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375132 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375155 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375172 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375189 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375204 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375222 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375236 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375250 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375266 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375280 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375297 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375317 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375335 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375349 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375363 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375382 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375398 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375434 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375451 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375466 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375486 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375502 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375519 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375539 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375555 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375572 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375587 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375610 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375626 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375640 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375652 4931 reconstruct.go:97] "Volume reconstruction finished" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.375663 4931 reconciler.go:26] "Reconciler: start to sync state" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.391332 4931 manager.go:324] Recovery completed Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.407630 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.409762 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.409824 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.409836 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.410631 4931 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.410649 4931 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.410680 4931 state_mem.go:36] "Initialized new in-memory state store" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.419138 4931 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.421269 4931 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.421333 4931 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.421362 4931 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.421639 4931 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.422667 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.422803 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.434237 4931 policy_none.go:49] "None policy: Start" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.435194 4931 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.435233 4931 state_mem.go:35] "Initializing new in-memory state store" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.443010 4931 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.511434 4931 manager.go:334] "Starting Device Plugin manager" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.511509 4931 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.511530 4931 server.go:79] "Starting device plugin registration server" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.512245 4931 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.512276 4931 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.512458 4931 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.512570 4931 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.512577 4931 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.520170 4931 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.522379 4931 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.522459 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.523458 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.523493 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.523507 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.523646 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.523831 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.523895 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.524491 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.524527 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.524546 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.524680 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.524814 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.524814 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.524908 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.524917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.524927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.525498 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.525535 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.525548 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.525664 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.525898 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.525924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.525962 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.525972 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.526099 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.526787 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.526872 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.526948 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.527369 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.527512 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.527556 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.528462 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.528507 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.528524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.531483 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.531515 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.531529 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.532571 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.532602 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.532616 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.532794 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.532832 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.534153 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.534226 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.534255 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.545310 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="400ms" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.581930 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.581980 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582009 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582036 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582059 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582080 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582101 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582130 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582153 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582184 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582214 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582293 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582365 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582416 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.582458 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.613347 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.614728 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.614842 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.614857 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.614912 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.615576 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684139 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684236 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684277 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684301 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684323 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684347 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684371 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684402 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684436 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684518 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684453 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684578 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684612 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684624 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684633 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684659 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684666 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684681 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684701 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684704 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684734 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684777 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684806 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684838 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684865 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684580 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684951 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.684994 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.685041 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.685088 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.816212 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.817932 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.818020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.818036 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.818075 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.818667 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.863368 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.869515 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.918912 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.922355 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ad111b0a78353d66066f27ad811f39f588c180162ee5db88a107c554de0031a6 WatchSource:0}: Error finding container ad111b0a78353d66066f27ad811f39f588c180162ee5db88a107c554de0031a6: Status 404 returned error can't find the container with id ad111b0a78353d66066f27ad811f39f588c180162ee5db88a107c554de0031a6 Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.932564 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: I1125 21:51:02.935548 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:02 crc kubenswrapper[4931]: E1125 21:51:02.946805 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="800ms" Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.955822 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-b53cd2bbb6a99a8bc9d8bd20a11753cfa65ec94ccb1dda586b12127b567e3f7f WatchSource:0}: Error finding container b53cd2bbb6a99a8bc9d8bd20a11753cfa65ec94ccb1dda586b12127b567e3f7f: Status 404 returned error can't find the container with id b53cd2bbb6a99a8bc9d8bd20a11753cfa65ec94ccb1dda586b12127b567e3f7f Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.972353 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-d29cd92d76d4a2ba19ef83ec1537fe7c3ff850463fef91ed15247b664df51362 WatchSource:0}: Error finding container d29cd92d76d4a2ba19ef83ec1537fe7c3ff850463fef91ed15247b664df51362: Status 404 returned error can't find the container with id d29cd92d76d4a2ba19ef83ec1537fe7c3ff850463fef91ed15247b664df51362 Nov 25 21:51:02 crc kubenswrapper[4931]: W1125 21:51:02.978570 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-54295ee77929aeeae726668a16ce3f46463ed5d5b05af6af691de5d34e3191b1 WatchSource:0}: Error finding container 54295ee77929aeeae726668a16ce3f46463ed5d5b05af6af691de5d34e3191b1: Status 404 returned error can't find the container with id 54295ee77929aeeae726668a16ce3f46463ed5d5b05af6af691de5d34e3191b1 Nov 25 21:51:03 crc kubenswrapper[4931]: W1125 21:51:03.158048 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:03 crc kubenswrapper[4931]: E1125 21:51:03.158140 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.225826 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.227488 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.227525 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.227536 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.227562 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 21:51:03 crc kubenswrapper[4931]: E1125 21:51:03.227967 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.339174 4931 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.342265 4931 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 03:02:23.507939234 +0000 UTC Nov 25 21:51:03 crc kubenswrapper[4931]: W1125 21:51:03.342525 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:03 crc kubenswrapper[4931]: E1125 21:51:03.342624 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.429334 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"54295ee77929aeeae726668a16ce3f46463ed5d5b05af6af691de5d34e3191b1"} Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.464917 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d29cd92d76d4a2ba19ef83ec1537fe7c3ff850463fef91ed15247b664df51362"} Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.466272 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b53cd2bbb6a99a8bc9d8bd20a11753cfa65ec94ccb1dda586b12127b567e3f7f"} Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.467411 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8ef4243025f8cd12890b6daad33daec66638b78633fe6288819b6a3c19ca2f08"} Nov 25 21:51:03 crc kubenswrapper[4931]: I1125 21:51:03.468257 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ad111b0a78353d66066f27ad811f39f588c180162ee5db88a107c554de0031a6"} Nov 25 21:51:03 crc kubenswrapper[4931]: W1125 21:51:03.510315 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:03 crc kubenswrapper[4931]: E1125 21:51:03.510445 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:03 crc kubenswrapper[4931]: W1125 21:51:03.622267 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:03 crc kubenswrapper[4931]: E1125 21:51:03.622367 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:03 crc kubenswrapper[4931]: E1125 21:51:03.748538 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="1.6s" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.028161 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.029541 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.029583 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.029601 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.029635 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 21:51:04 crc kubenswrapper[4931]: E1125 21:51:04.030258 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.339222 4931 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.343414 4931 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 16:06:41.166531676 +0000 UTC Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.343547 4931 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 738h15m36.822992312s for next certificate rotation Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.355667 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 25 21:51:04 crc kubenswrapper[4931]: E1125 21:51:04.357037 4931 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.473856 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c" exitCode=0 Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.474125 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.474722 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c"} Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.475497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.475557 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.475574 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.477953 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.478065 4931 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4" exitCode=0 Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.478189 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4"} Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.478259 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.479312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.479366 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.479314 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.479408 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.479379 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.479431 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.480370 4931 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a03f886d8278be42544d28620c0c8dd8ede5f2e578207a74377dc8ca11fec0a9" exitCode=0 Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.480414 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a03f886d8278be42544d28620c0c8dd8ede5f2e578207a74377dc8ca11fec0a9"} Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.480463 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.481525 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.481554 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.481572 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.483263 4931 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7" exitCode=0 Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.483367 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7"} Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.483390 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.485031 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.485194 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.485212 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.486129 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b"} Nov 25 21:51:04 crc kubenswrapper[4931]: I1125 21:51:04.486161 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.339247 4931 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:05 crc kubenswrapper[4931]: E1125 21:51:05.350943 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="3.2s" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.492869 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.492951 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.492964 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.493086 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.493991 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.494021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.494031 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.500518 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.500566 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.500683 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.501729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.501767 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.501778 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.507304 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.507351 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.507364 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.507375 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.509155 4931 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b" exitCode=0 Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.509206 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.509326 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.510395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.510423 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.510432 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.512897 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5190fb6f1f9bcffaf41a2be0ec4b9ef366220439b3a5d80061a910f393372c38"} Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.512977 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.513820 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.513841 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.513852 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:05 crc kubenswrapper[4931]: W1125 21:51:05.535232 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:05 crc kubenswrapper[4931]: E1125 21:51:05.535456 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.631023 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.632226 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.632266 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.632277 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:05 crc kubenswrapper[4931]: I1125 21:51:05.632300 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 21:51:05 crc kubenswrapper[4931]: E1125 21:51:05.632544 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 21:51:05 crc kubenswrapper[4931]: W1125 21:51:05.891803 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 21:51:05 crc kubenswrapper[4931]: E1125 21:51:05.891935 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.521441 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd"} Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.521832 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.524984 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.525043 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.525061 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.525430 4931 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba" exitCode=0 Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.525733 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.525951 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.526079 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba"} Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.526241 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.526323 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.526347 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.527932 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.527984 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528003 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528078 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528115 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528208 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528244 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528263 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528740 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:06 crc kubenswrapper[4931]: I1125 21:51:06.528782 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:07 crc kubenswrapper[4931]: I1125 21:51:07.530941 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976"} Nov 25 21:51:07 crc kubenswrapper[4931]: I1125 21:51:07.530982 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122"} Nov 25 21:51:07 crc kubenswrapper[4931]: I1125 21:51:07.530995 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c"} Nov 25 21:51:07 crc kubenswrapper[4931]: I1125 21:51:07.531033 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 21:51:07 crc kubenswrapper[4931]: I1125 21:51:07.531065 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:07 crc kubenswrapper[4931]: I1125 21:51:07.531784 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:07 crc kubenswrapper[4931]: I1125 21:51:07.531805 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:07 crc kubenswrapper[4931]: I1125 21:51:07.531815 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.520586 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.540267 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e"} Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.540352 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62"} Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.540601 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.542312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.542379 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.542424 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.586000 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.586323 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.587966 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.588032 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.588051 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.833389 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.835301 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.835374 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.835392 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:08 crc kubenswrapper[4931]: I1125 21:51:08.835431 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.193332 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.543987 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.544061 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.545951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.546017 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.546042 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.546235 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.546294 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.546311 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.843003 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.843255 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.845332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.845391 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:09 crc kubenswrapper[4931]: I1125 21:51:09.845415 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:10 crc kubenswrapper[4931]: I1125 21:51:10.576441 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:10 crc kubenswrapper[4931]: I1125 21:51:10.576748 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:10 crc kubenswrapper[4931]: I1125 21:51:10.578336 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:10 crc kubenswrapper[4931]: I1125 21:51:10.578408 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:10 crc kubenswrapper[4931]: I1125 21:51:10.578435 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:11 crc kubenswrapper[4931]: I1125 21:51:11.108115 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:11 crc kubenswrapper[4931]: I1125 21:51:11.108388 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:11 crc kubenswrapper[4931]: I1125 21:51:11.110019 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:11 crc kubenswrapper[4931]: I1125 21:51:11.110081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:11 crc kubenswrapper[4931]: I1125 21:51:11.110096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.220982 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.221195 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.222797 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.222917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.222955 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.229415 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:12 crc kubenswrapper[4931]: E1125 21:51:12.520467 4931 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.551843 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.552967 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.553007 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:12 crc kubenswrapper[4931]: I1125 21:51:12.553017 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.140838 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.141209 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.143024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.143127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.143151 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.169554 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.554412 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.555685 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.555744 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:13 crc kubenswrapper[4931]: I1125 21:51:13.555763 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:14 crc kubenswrapper[4931]: I1125 21:51:14.101319 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:14 crc kubenswrapper[4931]: I1125 21:51:14.557541 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:14 crc kubenswrapper[4931]: I1125 21:51:14.559782 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:14 crc kubenswrapper[4931]: I1125 21:51:14.559814 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:14 crc kubenswrapper[4931]: I1125 21:51:14.559823 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:14 crc kubenswrapper[4931]: I1125 21:51:14.566795 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:15 crc kubenswrapper[4931]: I1125 21:51:15.560305 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:15 crc kubenswrapper[4931]: I1125 21:51:15.561391 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:15 crc kubenswrapper[4931]: I1125 21:51:15.561442 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:15 crc kubenswrapper[4931]: I1125 21:51:15.561466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:16 crc kubenswrapper[4931]: E1125 21:51:16.296981 4931 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187b5e63ede4b69f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 21:51:02.331643551 +0000 UTC m=+0.688155710,LastTimestamp:2025-11-25 21:51:02.331643551 +0000 UTC m=+0.688155710,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 21:51:16 crc kubenswrapper[4931]: W1125 21:51:16.298095 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.298174 4931 trace.go:236] Trace[847853372]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 21:51:06.296) (total time: 10001ms): Nov 25 21:51:16 crc kubenswrapper[4931]: Trace[847853372]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (21:51:16.298) Nov 25 21:51:16 crc kubenswrapper[4931]: Trace[847853372]: [10.001243479s] [10.001243479s] END Nov 25 21:51:16 crc kubenswrapper[4931]: E1125 21:51:16.298196 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.339513 4931 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 21:51:16 crc kubenswrapper[4931]: W1125 21:51:16.529573 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.529669 4931 trace.go:236] Trace[859990650]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 21:51:06.528) (total time: 10001ms): Nov 25 21:51:16 crc kubenswrapper[4931]: Trace[859990650]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (21:51:16.529) Nov 25 21:51:16 crc kubenswrapper[4931]: Trace[859990650]: [10.001180848s] [10.001180848s] END Nov 25 21:51:16 crc kubenswrapper[4931]: E1125 21:51:16.529691 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.903961 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.904016 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.912023 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.912107 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.934103 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.934306 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.935651 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.935694 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:16 crc kubenswrapper[4931]: I1125 21:51:16.935711 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:17 crc kubenswrapper[4931]: I1125 21:51:17.102072 4931 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 21:51:17 crc kubenswrapper[4931]: I1125 21:51:17.102173 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.202227 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.202493 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.204142 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.204205 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.204224 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.209618 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.571115 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.571231 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.572561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.572614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:19 crc kubenswrapper[4931]: I1125 21:51:19.572637 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.004418 4931 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 21:51:21 crc kubenswrapper[4931]: E1125 21:51:21.907791 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.909077 4931 trace.go:236] Trace[1088694262]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 21:51:11.025) (total time: 10883ms): Nov 25 21:51:21 crc kubenswrapper[4931]: Trace[1088694262]: ---"Objects listed" error: 10883ms (21:51:21.908) Nov 25 21:51:21 crc kubenswrapper[4931]: Trace[1088694262]: [10.883245s] [10.883245s] END Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.909115 4931 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 21:51:21 crc kubenswrapper[4931]: E1125 21:51:21.916508 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.924144 4931 trace.go:236] Trace[495719324]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 21:51:10.412) (total time: 11511ms): Nov 25 21:51:21 crc kubenswrapper[4931]: Trace[495719324]: ---"Objects listed" error: 11511ms (21:51:21.924) Nov 25 21:51:21 crc kubenswrapper[4931]: Trace[495719324]: [11.51152943s] [11.51152943s] END Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.924174 4931 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.924409 4931 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.937187 4931 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.955579 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50958->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.955612 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48452->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.955648 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50958->192.168.126.11:17697: read: connection reset by peer" Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.955659 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48452->192.168.126.11:17697: read: connection reset by peer" Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.955979 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.956006 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.966959 4931 csr.go:261] certificate signing request csr-92l9w is approved, waiting to be issued Nov 25 21:51:21 crc kubenswrapper[4931]: I1125 21:51:21.987912 4931 csr.go:257] certificate signing request csr-92l9w is issued Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.157069 4931 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 25 21:51:22 crc kubenswrapper[4931]: W1125 21:51:22.157400 4931 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Node ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Nov 25 21:51:22 crc kubenswrapper[4931]: W1125 21:51:22.157426 4931 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.333733 4931 apiserver.go:52] "Watching apiserver" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.343793 4931 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.344200 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.344681 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.344778 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.344964 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.345147 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.345187 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.345357 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.347588 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.349486 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.349971 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.350278 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.350545 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.351730 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.351759 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.351779 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.353208 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.359794 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.359977 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.361663 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.415623 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.428234 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.443944 4931 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.445740 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.461439 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.474759 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.488633 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.499202 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.511667 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528277 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528326 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528348 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528373 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528408 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528427 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528462 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528484 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528506 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528527 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528546 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528591 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528610 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528633 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528701 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528723 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528765 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528788 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528807 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528826 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528847 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528866 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528907 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528941 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528961 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.528990 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529010 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529031 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529061 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529089 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529114 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529134 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529155 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529175 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529196 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529217 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529238 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529259 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529278 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529303 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529351 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529371 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529390 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529412 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529436 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529457 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529479 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529499 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529520 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529547 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529578 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529600 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529624 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529612 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529646 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529673 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529699 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529725 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529746 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529889 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529914 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529936 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.529971 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530022 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530044 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530064 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530087 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530108 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530127 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530152 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530176 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530199 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530222 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530243 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530263 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530285 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530306 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530330 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530338 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530351 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530636 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530680 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530720 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530760 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530795 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530830 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530863 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530933 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530976 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531008 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531041 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531073 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531121 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531154 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531186 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531232 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531267 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531300 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531334 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531370 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531419 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531460 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531501 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531540 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531576 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531614 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531670 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531703 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531743 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531786 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531820 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531854 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531917 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531959 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531996 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532038 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532071 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532110 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532144 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532187 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532468 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532516 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532578 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532622 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532657 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532702 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532743 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532781 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532823 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532867 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532929 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532964 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532998 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533033 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533068 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533100 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533136 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533177 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533211 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533249 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533285 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533319 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533356 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533398 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533433 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533467 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533504 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533538 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533573 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533612 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533649 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533684 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533720 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533754 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533797 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533830 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533864 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534059 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534095 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534146 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534187 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534221 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534264 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534299 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534335 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534370 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534405 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534439 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534486 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534528 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534565 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534604 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534639 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534680 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534715 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534751 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530378 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536091 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530429 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530462 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530590 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530701 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530735 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.530774 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531027 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531064 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531275 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531442 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531451 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531484 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531515 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536301 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531671 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536339 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531742 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531768 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531883 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531903 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.531968 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532068 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532215 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532355 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532417 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532460 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532548 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532670 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532751 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532799 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.532940 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533088 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533287 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533324 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533376 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533408 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533503 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.533585 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534209 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534536 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.534770 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:51:23.034752036 +0000 UTC m=+21.391264155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.534948 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.535511 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.535521 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.535647 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536018 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536432 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536567 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536686 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536734 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536761 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536790 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536824 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536853 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536855 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536903 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536947 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.536993 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539024 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539115 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539158 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539188 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539207 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539217 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539255 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539266 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539448 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539593 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539708 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539728 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.539790 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.540093 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.540219 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.541015 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.541568 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.541395 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.541644 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.541780 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.541854 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.541854 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.542359 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.542400 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.542824 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.543004 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.543283 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.543720 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.543853 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.543988 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.545571 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.545657 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.545713 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.545759 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.545803 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.545848 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.545925 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.545972 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.548930 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.549204 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.549954 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.550159 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.550825 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.550870 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.551181 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.551235 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.551571 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.552193 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.552821 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.553395 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.553513 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.553971 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.554032 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.554209 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.554660 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.554917 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.555058 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.555068 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.555121 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.555501 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.556131 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.556174 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.556370 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.556381 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.556754 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.556871 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.557274 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.557494 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.557580 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.558068 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.558673 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.558867 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.559116 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.559805 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.559932 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.560419 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.560626 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.561720 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.562085 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.562548 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.560536 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.563277 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.563289 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.563326 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.563722 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.563733 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.564181 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.564447 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.564728 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.564799 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.565052 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.565196 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566383 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566437 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566475 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566499 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566522 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566542 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.565507 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.562563 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.565764 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.565976 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.565999 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566200 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566221 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566752 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566350 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.567100 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.566566 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.567278 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.567344 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.567494 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.567944 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.568275 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.568388 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:23.068371175 +0000 UTC m=+21.424883294 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.568438 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:23.068428716 +0000 UTC m=+21.424940835 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.567961 4931 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.568053 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.568606 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.568834 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.568944 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.568895 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.569136 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.568688 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.568743 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.569579 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.569342 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.570345 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.570755 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.570771 4931 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.570789 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.570799 4931 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.570810 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.570973 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.570822 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571315 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571406 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571424 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571499 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571720 4931 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571761 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571795 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571919 4931 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571942 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571969 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.571987 4931 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572005 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572024 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572048 4931 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572070 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572090 4931 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572111 4931 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572136 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572156 4931 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572177 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572196 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572221 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572240 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572260 4931 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572286 4931 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572334 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572355 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572374 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572410 4931 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572429 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572448 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572467 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572490 4931 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572509 4931 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572527 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572550 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572568 4931 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572587 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572607 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572631 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572672 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572692 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572711 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572723 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572736 4931 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.572943 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.573656 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.579370 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.579867 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.579925 4931 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.579941 4931 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.579975 4931 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580000 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580014 4931 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580032 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580070 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580090 4931 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580105 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580139 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580159 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580171 4931 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580184 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580217 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580236 4931 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580250 4931 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580263 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580487 4931 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580509 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580525 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580560 4931 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580576 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580595 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580609 4931 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580643 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580663 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580677 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580691 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580725 4931 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580744 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580757 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580771 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580806 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580824 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580837 4931 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580850 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580898 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580916 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580930 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580943 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.580985 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581000 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581013 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581026 4931 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581065 4931 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581079 4931 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581093 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581109 4931 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581149 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581162 4931 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581176 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581214 4931 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581229 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581242 4931 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581260 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581300 4931 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581314 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581328 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581341 4931 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581377 4931 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581390 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581403 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581420 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581455 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581470 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581483 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581500 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581532 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581546 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581559 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581577 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581612 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581627 4931 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581642 4931 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581660 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581694 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581708 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581725 4931 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581737 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581772 4931 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581787 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581804 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581860 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581888 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581901 4931 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581918 4931 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581935 4931 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581949 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581963 4931 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581979 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.581991 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.582006 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.582022 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.587282 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.588515 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.590696 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.590852 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.590948 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.591023 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.591066 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:23.091033597 +0000 UTC m=+21.447545786 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.591030 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.591078 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.591126 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:22 crc kubenswrapper[4931]: E1125 21:51:22.591203 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:23.091176061 +0000 UTC m=+21.447688220 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.591529 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.591759 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.591858 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.594314 4931 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.594817 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.595695 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.596023 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.596502 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.596535 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.596996 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.597028 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.597305 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.598072 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.598218 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.598975 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.600013 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.600289 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.600519 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.605419 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.606168 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.606611 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.608206 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd" exitCode=255 Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.608251 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd"} Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.608923 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.613203 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.613941 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.618187 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.622078 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.622156 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.622467 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.622551 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.623101 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.623157 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.623299 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.622982 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.623372 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.623483 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.623647 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.624252 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.625150 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.625328 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.628063 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.628684 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.629456 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.629572 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.629635 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.630306 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.630454 4931 scope.go:117] "RemoveContainer" containerID="faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.632970 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.635248 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.635499 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.635563 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.644072 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.644144 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.654630 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.664021 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.676179 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.680383 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.682813 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.682937 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.682976 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683019 4931 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683058 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683072 4931 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683088 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683099 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683133 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683145 4931 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683156 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683168 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683180 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683215 4931 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683227 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683239 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683251 4931 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683263 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683296 4931 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683309 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683320 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683331 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683342 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683376 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683387 4931 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683399 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683411 4931 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683422 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683456 4931 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683467 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683478 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683491 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683503 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683539 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683551 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683566 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683580 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683592 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683625 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683637 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683649 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683071 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683663 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683696 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683709 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683721 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683732 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683743 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683776 4931 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683788 4931 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683800 4931 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683812 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683824 4931 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683853 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683864 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.683912 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.684815 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.693268 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.693599 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.700374 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.705593 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.988663 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-25 21:46:21 +0000 UTC, rotation deadline is 2026-10-10 03:18:13.638854085 +0000 UTC Nov 25 21:51:22 crc kubenswrapper[4931]: I1125 21:51:22.988735 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7637h26m50.650122993s for next certificate rotation Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.087474 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.087569 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.087617 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.087687 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:51:24.087666628 +0000 UTC m=+22.444178747 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.087731 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.087764 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.087812 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:24.087803511 +0000 UTC m=+22.444315630 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.087828 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:24.087821252 +0000 UTC m=+22.444333371 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.188741 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.188798 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.188974 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.189005 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.189020 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.189068 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:24.189052979 +0000 UTC m=+22.545565108 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.188974 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.189111 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.189143 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.189186 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:24.189174882 +0000 UTC m=+22.545686991 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.422135 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:23 crc kubenswrapper[4931]: E1125 21:51:23.422275 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.611504 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4c9ed20df84dfbc8928a825b905d214589eece20c81184ff5609d8a4b3b788da"} Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.613997 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186"} Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.614065 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9"} Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.614089 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"bf0e13fbd7fb544df8203a6fd8e3af3b6a5652eddd089a5c1f366a3845e9f933"} Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.616117 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a"} Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.616162 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9ce61020af5afb9124f31142120ffa1dfc9850d8c0e2f8f66bbbe2978f7c6776"} Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.618972 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.620399 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940"} Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.620913 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.653309 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.667751 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.687818 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.708810 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.726683 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.744321 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.767388 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.792468 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.818141 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.836474 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.863751 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.884363 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.922268 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:23 crc kubenswrapper[4931]: I1125 21:51:23.956264 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:23Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.095926 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.096006 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.096045 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.096101 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.096125 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:51:26.096099287 +0000 UTC m=+24.452611406 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.096159 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:26.096150849 +0000 UTC m=+24.452662968 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.096167 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.096252 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:26.096234671 +0000 UTC m=+24.452746840 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.105783 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.112508 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.121914 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.122461 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.136147 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.148579 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.157986 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.167844 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.194995 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.196517 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.196559 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.196663 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.196679 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.196688 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.196695 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.196715 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.196727 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.196751 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:26.196738039 +0000 UTC m=+24.553250158 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.196776 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:26.196760899 +0000 UTC m=+24.553273018 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.196925 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-xlsqn"] Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.197156 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xlsqn" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.198490 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.198892 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.198958 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.208646 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.221920 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.233122 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.247911 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.258615 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.271754 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.287199 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.297038 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckbgf\" (UniqueName: \"kubernetes.io/projected/1cd30323-d7c0-47e9-b280-2d71ef753965-kube-api-access-ckbgf\") pod \"node-resolver-xlsqn\" (UID: \"1cd30323-d7c0-47e9-b280-2d71ef753965\") " pod="openshift-dns/node-resolver-xlsqn" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.297084 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1cd30323-d7c0-47e9-b280-2d71ef753965-hosts-file\") pod \"node-resolver-xlsqn\" (UID: \"1cd30323-d7c0-47e9-b280-2d71ef753965\") " pod="openshift-dns/node-resolver-xlsqn" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.301856 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.313437 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.323509 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.398394 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckbgf\" (UniqueName: \"kubernetes.io/projected/1cd30323-d7c0-47e9-b280-2d71ef753965-kube-api-access-ckbgf\") pod \"node-resolver-xlsqn\" (UID: \"1cd30323-d7c0-47e9-b280-2d71ef753965\") " pod="openshift-dns/node-resolver-xlsqn" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.398482 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1cd30323-d7c0-47e9-b280-2d71ef753965-hosts-file\") pod \"node-resolver-xlsqn\" (UID: \"1cd30323-d7c0-47e9-b280-2d71ef753965\") " pod="openshift-dns/node-resolver-xlsqn" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.398625 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1cd30323-d7c0-47e9-b280-2d71ef753965-hosts-file\") pod \"node-resolver-xlsqn\" (UID: \"1cd30323-d7c0-47e9-b280-2d71ef753965\") " pod="openshift-dns/node-resolver-xlsqn" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.417606 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckbgf\" (UniqueName: \"kubernetes.io/projected/1cd30323-d7c0-47e9-b280-2d71ef753965-kube-api-access-ckbgf\") pod \"node-resolver-xlsqn\" (UID: \"1cd30323-d7c0-47e9-b280-2d71ef753965\") " pod="openshift-dns/node-resolver-xlsqn" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.422515 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.422554 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.422638 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.422742 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.426285 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.426796 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.427596 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.428181 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.428700 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.429191 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.429786 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.430408 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.431205 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.431861 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.432482 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.435605 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.436283 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.437395 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.438005 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.439097 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.439843 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.440366 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.441928 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.442721 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.443302 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.444738 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.445309 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.447292 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.447908 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.450717 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.451636 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.452236 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.453574 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.454179 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.455102 4931 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.455265 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.457693 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.458817 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.459319 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.462194 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.463768 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.464373 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.465566 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.466250 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.466707 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.467803 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.469710 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.470316 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.471118 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.471660 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.472562 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.473862 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.474895 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.475461 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.475930 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.476916 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.477502 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.478138 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.506406 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xlsqn" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.590531 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-djjhg"] Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.591107 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.592470 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-fcwz5"] Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.593034 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.595458 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-xltbl"] Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.595649 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.595782 4931 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.595819 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.596159 4931 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.596176 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.596440 4931 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.596468 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.596913 4931 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.596935 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.597129 4931 reflector.go:561] object-"openshift-multus"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.597148 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.597384 4931 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.597404 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.597775 4931 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.597795 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.597774 4931 reflector.go:561] object-"openshift-multus"/"default-cni-sysctl-allowlist": failed to list *v1.ConfigMap: configmaps "default-cni-sysctl-allowlist" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.597814 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"default-cni-sysctl-allowlist\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.598750 4931 reflector.go:561] object-"openshift-machine-config-operator"/"proxy-tls": failed to list *v1.Secret: secrets "proxy-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.598790 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.598797 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"proxy-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"proxy-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: W1125 21:51:24.598926 4931 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.598940 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 21:51:24 crc kubenswrapper[4931]: E1125 21:51:24.598946 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.615057 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.631281 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xlsqn" event={"ID":"1cd30323-d7c0-47e9-b280-2d71ef753965","Type":"ContainerStarted","Data":"8d274c1b767469acac0e41f77b5cc5038573eb02c47a105a685e6d69859dfe40"} Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.633328 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.646902 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.661168 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.673910 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.684719 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.697258 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.699912 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-conf-dir\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.699965 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e22448de-5122-43be-853a-081347c13cf2-proxy-tls\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.699989 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-var-lib-cni-multus\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700021 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700038 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e22448de-5122-43be-853a-081347c13cf2-mcd-auth-proxy-config\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700086 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-run-multus-certs\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700122 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9btrl\" (UniqueName: \"kubernetes.io/projected/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-kube-api-access-9btrl\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700147 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-cni-dir\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700182 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-cnibin\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700202 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-run-k8s-cni-cncf-io\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700227 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/53c90ad5-9f43-44c7-b122-0303c788f562-cni-binary-copy\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700247 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-socket-dir-parent\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700267 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-run-netns\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700289 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-var-lib-kubelet\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700338 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-os-release\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700450 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-system-cni-dir\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700472 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-daemon-config\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700503 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-cnibin\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700525 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh64z\" (UniqueName: \"kubernetes.io/projected/e22448de-5122-43be-853a-081347c13cf2-kube-api-access-xh64z\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700549 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-var-lib-cni-bin\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700570 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-cni-binary-copy\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700664 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/53c90ad5-9f43-44c7-b122-0303c788f562-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700719 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcl58\" (UniqueName: \"kubernetes.io/projected/53c90ad5-9f43-44c7-b122-0303c788f562-kube-api-access-jcl58\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700757 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e22448de-5122-43be-853a-081347c13cf2-rootfs\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700783 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-system-cni-dir\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700817 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-os-release\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700848 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-hostroot\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.700870 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-etc-kubernetes\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.714104 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.731502 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.745845 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.765182 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.779605 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.791017 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.800853 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801364 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-var-lib-cni-multus\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801405 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801427 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e22448de-5122-43be-853a-081347c13cf2-proxy-tls\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801450 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e22448de-5122-43be-853a-081347c13cf2-mcd-auth-proxy-config\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801474 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9btrl\" (UniqueName: \"kubernetes.io/projected/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-kube-api-access-9btrl\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801494 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-cni-dir\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801515 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-run-multus-certs\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801537 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-cnibin\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801558 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-run-k8s-cni-cncf-io\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801579 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/53c90ad5-9f43-44c7-b122-0303c788f562-cni-binary-copy\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801603 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-socket-dir-parent\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801621 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-run-netns\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801640 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-var-lib-kubelet\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801668 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-os-release\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801687 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-system-cni-dir\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801706 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-daemon-config\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801741 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-cnibin\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801757 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-cni-dir\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801811 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-run-multus-certs\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801866 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-cnibin\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801914 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-run-k8s-cni-cncf-io\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801981 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-var-lib-kubelet\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801761 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh64z\" (UniqueName: \"kubernetes.io/projected/e22448de-5122-43be-853a-081347c13cf2-kube-api-access-xh64z\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.801573 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-var-lib-cni-multus\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802034 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-cni-binary-copy\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802042 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-socket-dir-parent\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802057 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-var-lib-cni-bin\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802071 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-run-netns\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802068 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802102 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/53c90ad5-9f43-44c7-b122-0303c788f562-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802117 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-os-release\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802124 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-system-cni-dir\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802151 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-host-var-lib-cni-bin\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802164 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-cnibin\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802225 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcl58\" (UniqueName: \"kubernetes.io/projected/53c90ad5-9f43-44c7-b122-0303c788f562-kube-api-access-jcl58\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802251 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e22448de-5122-43be-853a-081347c13cf2-rootfs\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802275 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-system-cni-dir\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802297 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-os-release\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802316 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e22448de-5122-43be-853a-081347c13cf2-rootfs\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802324 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-etc-kubernetes\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802347 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-hostroot\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802371 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-conf-dir\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802431 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-conf-dir\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802462 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53c90ad5-9f43-44c7-b122-0303c788f562-system-cni-dir\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802468 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-etc-kubernetes\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802491 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-hostroot\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.802510 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-os-release\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.811277 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.822499 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.834945 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.849064 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.864724 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.877678 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.894089 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.906457 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.974689 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jjhc2"] Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.975725 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.980851 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.981153 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.981391 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.981481 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.981770 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.981826 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.983255 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 21:51:24 crc kubenswrapper[4931]: I1125 21:51:24.993633 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:24Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003246 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-netns\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003287 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-kubelet\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003344 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-openvswitch\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003367 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-ovn\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003387 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nntdj\" (UniqueName: \"kubernetes.io/projected/b6be6a2b-7522-4af3-aa00-68fa970fd81c-kube-api-access-nntdj\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003409 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-ovn-kubernetes\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003431 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-systemd\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003453 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-var-lib-openvswitch\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003472 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-node-log\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003488 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-config\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003528 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-systemd-units\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003550 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-netd\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003594 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-log-socket\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003617 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-bin\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003646 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-script-lib\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003699 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-slash\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003722 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003745 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-env-overrides\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003764 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-etc-openvswitch\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.003786 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovn-node-metrics-cert\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.008690 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.030948 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.049546 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.061501 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.077945 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.092759 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.104846 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-netns\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105018 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-netns\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105187 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-kubelet\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105055 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-kubelet\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105317 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-ovn\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105345 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nntdj\" (UniqueName: \"kubernetes.io/projected/b6be6a2b-7522-4af3-aa00-68fa970fd81c-kube-api-access-nntdj\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105378 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-openvswitch\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105405 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-ovn-kubernetes\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105426 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-systemd\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105448 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-var-lib-openvswitch\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105450 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-ovn\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105492 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-node-log\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105467 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-node-log\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105522 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-openvswitch\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105553 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-ovn-kubernetes\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105546 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-config\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105591 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-systemd-units\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105653 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-netd\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105699 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-log-socket\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105718 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-bin\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105749 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-script-lib\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105802 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-slash\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105824 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105847 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-env-overrides\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105893 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-etc-openvswitch\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105913 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovn-node-metrics-cert\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105913 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-systemd\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105949 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-var-lib-openvswitch\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105980 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-netd\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.106015 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.106063 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-log-socket\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.106108 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-bin\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.105843 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-systemd-units\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.106272 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-etc-openvswitch\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.106373 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-slash\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.106761 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-config\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.107451 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-env-overrides\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.108157 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-script-lib\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.110078 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.114574 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovn-node-metrics-cert\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.124508 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.128574 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nntdj\" (UniqueName: \"kubernetes.io/projected/b6be6a2b-7522-4af3-aa00-68fa970fd81c-kube-api-access-nntdj\") pod \"ovnkube-node-jjhc2\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.144512 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.168202 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.181935 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.193546 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.293607 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:25 crc kubenswrapper[4931]: W1125 21:51:25.309160 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6be6a2b_7522_4af3_aa00_68fa970fd81c.slice/crio-5c01cc3149ce9ff29e4717d833639e7977a6755801166e0c7a511acf156515b0 WatchSource:0}: Error finding container 5c01cc3149ce9ff29e4717d833639e7977a6755801166e0c7a511acf156515b0: Status 404 returned error can't find the container with id 5c01cc3149ce9ff29e4717d833639e7977a6755801166e0c7a511acf156515b0 Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.421915 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.422055 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.422865 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.425382 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.434469 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh64z\" (UniqueName: \"kubernetes.io/projected/e22448de-5122-43be-853a-081347c13cf2-kube-api-access-xh64z\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.435995 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e22448de-5122-43be-853a-081347c13cf2-proxy-tls\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.548255 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.552714 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/53c90ad5-9f43-44c7-b122-0303c788f562-cni-binary-copy\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.553263 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-cni-binary-copy\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.636248 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98" exitCode=0 Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.636368 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98"} Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.636411 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"5c01cc3149ce9ff29e4717d833639e7977a6755801166e0c7a511acf156515b0"} Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.639067 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927"} Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.641096 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xlsqn" event={"ID":"1cd30323-d7c0-47e9-b280-2d71ef753965","Type":"ContainerStarted","Data":"6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487"} Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.658934 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.680748 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.703821 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.722818 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.743278 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.760140 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.780375 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.790470 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.792352 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.793207 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-multus-daemon-config\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.802328 4931 configmap.go:193] Couldn't get configMap openshift-multus/default-cni-sysctl-allowlist: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.802391 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/53c90ad5-9f43-44c7-b122-0303c788f562-cni-sysctl-allowlist podName:53c90ad5-9f43-44c7-b122-0303c788f562 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:26.302375258 +0000 UTC m=+24.658887387 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-sysctl-allowlist" (UniqueName: "kubernetes.io/configmap/53c90ad5-9f43-44c7-b122-0303c788f562-cni-sysctl-allowlist") pod "multus-additional-cni-plugins-fcwz5" (UID: "53c90ad5-9f43-44c7-b122-0303c788f562") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.802430 4931 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.802457 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e22448de-5122-43be-853a-081347c13cf2-mcd-auth-proxy-config podName:e22448de-5122-43be-853a-081347c13cf2 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:26.30244971 +0000 UTC m=+24.658961839 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "mcd-auth-proxy-config" (UniqueName: "kubernetes.io/configmap/e22448de-5122-43be-853a-081347c13cf2-mcd-auth-proxy-config") pod "machine-config-daemon-djjhg" (UID: "e22448de-5122-43be-853a-081347c13cf2") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.807550 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.810095 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.817858 4931 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.817896 4931 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.817930 4931 projected.go:194] Error preparing data for projected volume kube-api-access-jcl58 for pod openshift-multus/multus-additional-cni-plugins-fcwz5: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.817941 4931 projected.go:194] Error preparing data for projected volume kube-api-access-9btrl for pod openshift-multus/multus-xltbl: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.818008 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/53c90ad5-9f43-44c7-b122-0303c788f562-kube-api-access-jcl58 podName:53c90ad5-9f43-44c7-b122-0303c788f562 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:26.31798802 +0000 UTC m=+24.674500149 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-jcl58" (UniqueName: "kubernetes.io/projected/53c90ad5-9f43-44c7-b122-0303c788f562-kube-api-access-jcl58") pod "multus-additional-cni-plugins-fcwz5" (UID: "53c90ad5-9f43-44c7-b122-0303c788f562") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: E1125 21:51:25.818030 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-kube-api-access-9btrl podName:ef281bd7-e1b1-41cd-b0a2-fa026829cf8f nodeName:}" failed. No retries permitted until 2025-11-25 21:51:26.318022211 +0000 UTC m=+24.674534340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-9btrl" (UniqueName: "kubernetes.io/projected/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-kube-api-access-9btrl") pod "multus-xltbl" (UID: "ef281bd7-e1b1-41cd-b0a2-fa026829cf8f") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.830537 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.845409 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.862748 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.875716 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.903317 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.921092 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.937653 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.949139 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.954714 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.956927 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.961710 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.973074 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.986371 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:25 crc kubenswrapper[4931]: I1125 21:51:25.998190 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:25Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.014248 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:26Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.027705 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:26Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.037596 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:26Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.054566 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:26Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.061357 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.068343 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.076179 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:26Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.095483 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.116113 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.116255 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.116317 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:51:30.116286307 +0000 UTC m=+28.472798436 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.116348 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.116382 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.116405 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:30.11638908 +0000 UTC m=+28.472901209 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.116560 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.116625 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:30.116610965 +0000 UTC m=+28.473123094 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.217118 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.217180 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.217321 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.217332 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.217349 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.217361 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.217361 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.217375 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.217444 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:30.217422431 +0000 UTC m=+28.573934570 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.217470 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:30.217458162 +0000 UTC m=+28.573970301 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.318332 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e22448de-5122-43be-853a-081347c13cf2-mcd-auth-proxy-config\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.318399 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9btrl\" (UniqueName: \"kubernetes.io/projected/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-kube-api-access-9btrl\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.318503 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/53c90ad5-9f43-44c7-b122-0303c788f562-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.318544 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcl58\" (UniqueName: \"kubernetes.io/projected/53c90ad5-9f43-44c7-b122-0303c788f562-kube-api-access-jcl58\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.319301 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e22448de-5122-43be-853a-081347c13cf2-mcd-auth-proxy-config\") pod \"machine-config-daemon-djjhg\" (UID: \"e22448de-5122-43be-853a-081347c13cf2\") " pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.319835 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/53c90ad5-9f43-44c7-b122-0303c788f562-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.322228 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcl58\" (UniqueName: \"kubernetes.io/projected/53c90ad5-9f43-44c7-b122-0303c788f562-kube-api-access-jcl58\") pod \"multus-additional-cni-plugins-fcwz5\" (UID: \"53c90ad5-9f43-44c7-b122-0303c788f562\") " pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.325052 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9btrl\" (UniqueName: \"kubernetes.io/projected/ef281bd7-e1b1-41cd-b0a2-fa026829cf8f-kube-api-access-9btrl\") pod \"multus-xltbl\" (UID: \"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\") " pod="openshift-multus/multus-xltbl" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.403211 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.412418 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.417678 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xltbl" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.422266 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.422332 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.422478 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:26 crc kubenswrapper[4931]: E1125 21:51:26.422644 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:26 crc kubenswrapper[4931]: W1125 21:51:26.431869 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode22448de_5122_43be_853a_081347c13cf2.slice/crio-bedd3272bd418817f652374ad6aaaceb3b7e61ea7c4c76dc5b1a67ce0dfa24f2 WatchSource:0}: Error finding container bedd3272bd418817f652374ad6aaaceb3b7e61ea7c4c76dc5b1a67ce0dfa24f2: Status 404 returned error can't find the container with id bedd3272bd418817f652374ad6aaaceb3b7e61ea7c4c76dc5b1a67ce0dfa24f2 Nov 25 21:51:26 crc kubenswrapper[4931]: W1125 21:51:26.433616 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53c90ad5_9f43_44c7_b122_0303c788f562.slice/crio-a6134fbe8abb26f7d62b14f27904c1972ed468b8fe0534f346c9e3cb3330dc72 WatchSource:0}: Error finding container a6134fbe8abb26f7d62b14f27904c1972ed468b8fe0534f346c9e3cb3330dc72: Status 404 returned error can't find the container with id a6134fbe8abb26f7d62b14f27904c1972ed468b8fe0534f346c9e3cb3330dc72 Nov 25 21:51:26 crc kubenswrapper[4931]: W1125 21:51:26.439535 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef281bd7_e1b1_41cd_b0a2_fa026829cf8f.slice/crio-bb313ee5b400023fbb5b3ed5f318e9b5fd9bf597a7844a0c22cd1f5bd484dba6 WatchSource:0}: Error finding container bb313ee5b400023fbb5b3ed5f318e9b5fd9bf597a7844a0c22cd1f5bd484dba6: Status 404 returned error can't find the container with id bb313ee5b400023fbb5b3ed5f318e9b5fd9bf597a7844a0c22cd1f5bd484dba6 Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.651351 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca"} Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.651399 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d"} Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.651412 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228"} Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.651425 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d"} Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.651437 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b"} Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.651471 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c"} Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.652601 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"bedd3272bd418817f652374ad6aaaceb3b7e61ea7c4c76dc5b1a67ce0dfa24f2"} Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.653676 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xltbl" event={"ID":"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f","Type":"ContainerStarted","Data":"bb313ee5b400023fbb5b3ed5f318e9b5fd9bf597a7844a0c22cd1f5bd484dba6"} Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.654470 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerStarted","Data":"a6134fbe8abb26f7d62b14f27904c1972ed468b8fe0534f346c9e3cb3330dc72"} Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.965159 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.976159 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.979162 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:26Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:26 crc kubenswrapper[4931]: I1125 21:51:26.980182 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.000763 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:26Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.032749 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.066056 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.079885 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.091982 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.101835 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.112179 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.126941 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.140634 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.152586 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.163454 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.173958 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.187173 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.199119 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.212559 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.220225 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-dfn4g"] Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.220810 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.222492 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.223033 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.223401 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.226455 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.227595 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.242251 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.265963 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.279865 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.303200 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.315293 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.325287 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.327294 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2e774b7-8bb0-4625-a50d-f69effe7ccad-host\") pod \"node-ca-dfn4g\" (UID: \"c2e774b7-8bb0-4625-a50d-f69effe7ccad\") " pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.327358 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbk6w\" (UniqueName: \"kubernetes.io/projected/c2e774b7-8bb0-4625-a50d-f69effe7ccad-kube-api-access-fbk6w\") pod \"node-ca-dfn4g\" (UID: \"c2e774b7-8bb0-4625-a50d-f69effe7ccad\") " pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.327420 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c2e774b7-8bb0-4625-a50d-f69effe7ccad-serviceca\") pod \"node-ca-dfn4g\" (UID: \"c2e774b7-8bb0-4625-a50d-f69effe7ccad\") " pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.340937 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.353479 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.368458 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.385161 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.414455 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.421568 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:27 crc kubenswrapper[4931]: E1125 21:51:27.421732 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.428314 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbk6w\" (UniqueName: \"kubernetes.io/projected/c2e774b7-8bb0-4625-a50d-f69effe7ccad-kube-api-access-fbk6w\") pod \"node-ca-dfn4g\" (UID: \"c2e774b7-8bb0-4625-a50d-f69effe7ccad\") " pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.428446 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c2e774b7-8bb0-4625-a50d-f69effe7ccad-serviceca\") pod \"node-ca-dfn4g\" (UID: \"c2e774b7-8bb0-4625-a50d-f69effe7ccad\") " pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.428495 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2e774b7-8bb0-4625-a50d-f69effe7ccad-host\") pod \"node-ca-dfn4g\" (UID: \"c2e774b7-8bb0-4625-a50d-f69effe7ccad\") " pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.428616 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2e774b7-8bb0-4625-a50d-f69effe7ccad-host\") pod \"node-ca-dfn4g\" (UID: \"c2e774b7-8bb0-4625-a50d-f69effe7ccad\") " pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.430424 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c2e774b7-8bb0-4625-a50d-f69effe7ccad-serviceca\") pod \"node-ca-dfn4g\" (UID: \"c2e774b7-8bb0-4625-a50d-f69effe7ccad\") " pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.436441 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.453289 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.463448 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbk6w\" (UniqueName: \"kubernetes.io/projected/c2e774b7-8bb0-4625-a50d-f69effe7ccad-kube-api-access-fbk6w\") pod \"node-ca-dfn4g\" (UID: \"c2e774b7-8bb0-4625-a50d-f69effe7ccad\") " pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.476653 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.494177 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.516752 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.534347 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dfn4g" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.549333 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: W1125 21:51:27.557065 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2e774b7_8bb0_4625_a50d_f69effe7ccad.slice/crio-25af0f8f0d12b94c91ee5ba2976455767c0975afd47bd44c1fe391f5825918b3 WatchSource:0}: Error finding container 25af0f8f0d12b94c91ee5ba2976455767c0975afd47bd44c1fe391f5825918b3: Status 404 returned error can't find the container with id 25af0f8f0d12b94c91ee5ba2976455767c0975afd47bd44c1fe391f5825918b3 Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.573517 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.595578 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.625797 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.659751 4931 generic.go:334] "Generic (PLEG): container finished" podID="53c90ad5-9f43-44c7-b122-0303c788f562" containerID="a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120" exitCode=0 Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.659818 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerDied","Data":"a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120"} Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.668066 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dfn4g" event={"ID":"c2e774b7-8bb0-4625-a50d-f69effe7ccad","Type":"ContainerStarted","Data":"25af0f8f0d12b94c91ee5ba2976455767c0975afd47bd44c1fe391f5825918b3"} Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.669858 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9"} Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.669955 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916"} Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.671128 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xltbl" event={"ID":"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f","Type":"ContainerStarted","Data":"fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4"} Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.675417 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.702867 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.742087 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.788594 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.822025 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.871023 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.902766 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:27 crc kubenswrapper[4931]: I1125 21:51:27.950617 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.000065 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.030969 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.073355 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.109436 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.147357 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.208768 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.232811 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.269750 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.313767 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.316801 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.322553 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.323349 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.323398 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.323590 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.347191 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.397857 4931 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.398138 4931 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.399427 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.399474 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.399487 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.399504 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.399516 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: E1125 21:51:28.418079 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.422277 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.422356 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:28 crc kubenswrapper[4931]: E1125 21:51:28.422472 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:28 crc kubenswrapper[4931]: E1125 21:51:28.422548 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.423965 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.423999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.424012 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.424029 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.424041 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.432530 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: E1125 21:51:28.442297 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.447920 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.447999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.448023 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.448416 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.448435 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: E1125 21:51:28.468408 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.471744 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.473814 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.473916 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.473935 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.473959 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.473976 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: E1125 21:51:28.497485 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.502690 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.502732 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.502741 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.502758 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.502767 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: E1125 21:51:28.529052 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: E1125 21:51:28.529230 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.530690 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.530725 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.530735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.530749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.530759 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.632332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.632364 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.632374 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.632387 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.632397 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.677014 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerStarted","Data":"3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a"} Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.678745 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dfn4g" event={"ID":"c2e774b7-8bb0-4625-a50d-f69effe7ccad","Type":"ContainerStarted","Data":"cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be"} Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.684394 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34"} Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.696780 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.709333 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.733950 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.735209 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.735236 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.735245 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.735259 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.735268 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.744689 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.757209 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.768569 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.777117 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.787967 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.832952 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.838119 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.838150 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.838160 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.838172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.838181 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.863605 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.914895 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.941979 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.942025 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.942042 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.942064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.942081 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:28Z","lastTransitionTime":"2025-11-25T21:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.953162 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:28 crc kubenswrapper[4931]: I1125 21:51:28.991499 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.027844 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.045738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.045775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.045791 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.045815 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.045829 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:29Z","lastTransitionTime":"2025-11-25T21:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.065331 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.119105 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.147680 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.148216 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.148274 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.148299 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.148330 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.148354 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:29Z","lastTransitionTime":"2025-11-25T21:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.184746 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.234759 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.251540 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.251680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.251708 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.251738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.251761 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:29Z","lastTransitionTime":"2025-11-25T21:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.278850 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.311658 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.373569 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.373653 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.373682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.373717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.373742 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:29Z","lastTransitionTime":"2025-11-25T21:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.383548 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.423277 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.424026 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:29 crc kubenswrapper[4931]: E1125 21:51:29.426308 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.440860 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.464283 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.476910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.476960 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.476977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.477002 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.477023 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:29Z","lastTransitionTime":"2025-11-25T21:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.504378 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.547486 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.580021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.580462 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.580486 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.580514 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.580539 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:29Z","lastTransitionTime":"2025-11-25T21:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.587573 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.625388 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.682683 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.692464 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.692522 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.692539 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.692562 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.692579 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:29Z","lastTransitionTime":"2025-11-25T21:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.697756 4931 generic.go:334] "Generic (PLEG): container finished" podID="53c90ad5-9f43-44c7-b122-0303c788f562" containerID="3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a" exitCode=0 Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.697930 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerDied","Data":"3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.721410 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.748674 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.791824 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.796233 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.796264 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.796275 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.796291 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.796304 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:29Z","lastTransitionTime":"2025-11-25T21:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.829241 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.867366 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.898845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.898908 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.898922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.898938 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.898949 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:29Z","lastTransitionTime":"2025-11-25T21:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.911028 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.950851 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:29 crc kubenswrapper[4931]: I1125 21:51:29.991176 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:29Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.002053 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.002117 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.002141 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.002173 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.002196 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.035807 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.071778 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.105214 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.105276 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.105294 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.105319 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.105336 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.110826 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.153646 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.174218 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.174427 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:51:38.174381695 +0000 UTC m=+36.530893854 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.174702 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.174864 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.175016 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:38.174990132 +0000 UTC m=+36.531502341 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.175170 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.175260 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:38.175237178 +0000 UTC m=+36.531749337 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.175436 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.200195 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.208534 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.208590 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.208617 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.208648 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.208670 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.240014 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.269455 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.277027 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.277342 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.277403 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.277473 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.277643 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.277735 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:38.277696919 +0000 UTC m=+36.634209098 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.277855 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.277925 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.277944 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.278023 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:38.278000957 +0000 UTC m=+36.634513146 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.311843 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.311962 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.311992 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.312025 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.312049 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.415218 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.415293 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.415317 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.415349 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.415371 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.421804 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.421932 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.422070 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:30 crc kubenswrapper[4931]: E1125 21:51:30.422198 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.518549 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.518613 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.518630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.518655 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.518674 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.521275 4931 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.621141 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.621179 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.621188 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.621202 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.621211 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.703153 4931 generic.go:334] "Generic (PLEG): container finished" podID="53c90ad5-9f43-44c7-b122-0303c788f562" containerID="4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8" exitCode=0 Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.703207 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerDied","Data":"4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.724225 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.724524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.724536 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.724551 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.724562 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.734561 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.750107 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.768577 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.786176 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.804732 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.818487 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.827181 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.827216 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.827224 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.827240 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.827249 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.833637 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.854020 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.865140 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.875685 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.891352 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.905635 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.917140 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.930046 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.930093 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.930103 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.930120 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.930129 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:30Z","lastTransitionTime":"2025-11-25T21:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.930769 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:30 crc kubenswrapper[4931]: I1125 21:51:30.952814 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:30Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.032668 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.032711 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.032943 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.032969 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.032981 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.135237 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.135272 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.135281 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.135293 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.135302 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.238822 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.238945 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.238967 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.238990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.239007 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.341814 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.341859 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.341895 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.341919 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.341939 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.422431 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:31 crc kubenswrapper[4931]: E1125 21:51:31.422571 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.445001 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.445048 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.445060 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.445078 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.445089 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.547996 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.548071 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.548090 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.548114 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.548134 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.651746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.652211 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.652230 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.652256 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.652277 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.711364 4931 generic.go:334] "Generic (PLEG): container finished" podID="53c90ad5-9f43-44c7-b122-0303c788f562" containerID="b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338" exitCode=0 Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.711480 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerDied","Data":"b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.718461 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.718945 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.719008 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.736434 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.754783 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.755971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.756087 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.756114 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.756144 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.756166 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.759462 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.763295 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.772334 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.789039 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.808602 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.830076 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.847255 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.859117 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.859157 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.859171 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.859192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.859207 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.871383 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.892812 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.911193 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.922424 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.938385 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.956097 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.961278 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.961323 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.961336 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.961363 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.961379 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:31Z","lastTransitionTime":"2025-11-25T21:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:31 crc kubenswrapper[4931]: I1125 21:51:31.975081 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:31Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.003698 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.019074 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.036393 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.049708 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.064568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.064616 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.064630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.064655 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.064681 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:32Z","lastTransitionTime":"2025-11-25T21:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.066530 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.087925 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.106959 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.140914 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.156849 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.167152 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.167212 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.167231 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.167256 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.167274 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:32Z","lastTransitionTime":"2025-11-25T21:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.176264 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.194350 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.209442 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.222966 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.244237 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.258503 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.269991 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.270192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.270284 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.270374 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.270484 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:32Z","lastTransitionTime":"2025-11-25T21:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.272188 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.372676 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.373034 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.373196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.373322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.373502 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:32Z","lastTransitionTime":"2025-11-25T21:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.421980 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:32 crc kubenswrapper[4931]: E1125 21:51:32.422179 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.422699 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:32 crc kubenswrapper[4931]: E1125 21:51:32.424065 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.457163 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.476744 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.476813 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.476832 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.476865 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.476916 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:32Z","lastTransitionTime":"2025-11-25T21:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.478169 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.500558 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.521119 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.548941 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.571568 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.583172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.583251 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.583274 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.583305 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.583325 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:32Z","lastTransitionTime":"2025-11-25T21:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.587613 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.602082 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.614086 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.630758 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.651155 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.685420 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.688026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.688101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.688120 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.688146 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.688165 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:32Z","lastTransitionTime":"2025-11-25T21:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.703468 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.719971 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.725846 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerStarted","Data":"ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.726015 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.734967 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.753264 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.771733 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.791400 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.791467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.791479 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.791508 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.791523 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:32Z","lastTransitionTime":"2025-11-25T21:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.810443 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.848283 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.897172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.897225 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.897241 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.897261 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.897275 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:32Z","lastTransitionTime":"2025-11-25T21:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.898434 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.933202 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:32 crc kubenswrapper[4931]: I1125 21:51:32.969229 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:32Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.000667 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.000714 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.000726 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.000746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.000759 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.012175 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.049892 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.093391 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.103059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.103095 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.103103 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.103117 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.103126 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.126253 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.170234 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.205725 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.205767 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.205780 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.205797 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.205809 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.215525 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.246637 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.296813 4931 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.309116 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.309164 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.309178 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.309196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.309209 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.320983 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.411808 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.411844 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.411854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.411869 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.411897 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.422383 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:33 crc kubenswrapper[4931]: E1125 21:51:33.422505 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.514699 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.514737 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.514750 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.514769 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.514782 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.617131 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.617164 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.617173 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.617185 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.617194 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.719953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.719998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.720009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.720027 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.720039 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.731013 4931 generic.go:334] "Generic (PLEG): container finished" podID="53c90ad5-9f43-44c7-b122-0303c788f562" containerID="ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8" exitCode=0 Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.731078 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerDied","Data":"ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.731241 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.743118 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.760991 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.790077 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.802607 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.817669 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.822891 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.822928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.822938 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.822954 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.822964 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.831083 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.855487 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.876182 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.890223 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.912335 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.925067 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.925101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.925109 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.925122 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.925133 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:33Z","lastTransitionTime":"2025-11-25T21:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.929801 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.941778 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.956744 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.974804 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:33 crc kubenswrapper[4931]: I1125 21:51:33.985892 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:33Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.027579 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.027621 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.027630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.027644 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.027654 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.130746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.130799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.130811 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.130829 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.130841 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.233395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.233491 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.233949 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.233998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.234013 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.336153 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.336190 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.336198 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.336211 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.336221 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.422581 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.422673 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:34 crc kubenswrapper[4931]: E1125 21:51:34.422750 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:34 crc kubenswrapper[4931]: E1125 21:51:34.422781 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.438665 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.438729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.438750 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.438775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.438795 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.541410 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.541477 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.541496 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.541520 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.541538 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.644521 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.644550 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.644559 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.644570 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.644579 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.738197 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/0.log" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.743123 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352" exitCode=1 Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.743179 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.744730 4931 scope.go:117] "RemoveContainer" containerID="508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.748087 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.748186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.748267 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.748441 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.748528 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.754359 4931 generic.go:334] "Generic (PLEG): container finished" podID="53c90ad5-9f43-44c7-b122-0303c788f562" containerID="edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75" exitCode=0 Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.754423 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerDied","Data":"edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.759920 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.783796 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.804408 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"message\\\":\\\"1125 21:51:33.767213 6140 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:33.767153 6140 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 21:51:33.767233 6140 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 21:51:33.767315 6140 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 21:51:33.767390 6140 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 21:51:33.767464 6140 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 21:51:33.767466 6140 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:33.767481 6140 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:33.767496 6140 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:33.767502 6140 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:33.767562 6140 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 21:51:33.767960 6140 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:33.768016 6140 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 21:51:33.768063 6140 factory.go:656] Stopping watch factory\\\\nI1125 21:51:33.768081 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1125 21:51:33.768105 6140 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.820571 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.839405 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.852914 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.852959 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.852971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.852990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.853021 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.856812 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.887587 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.907345 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.920149 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.936729 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.955476 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.955437 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.955501 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.955677 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.955691 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.955700 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:34Z","lastTransitionTime":"2025-11-25T21:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.971091 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:34 crc kubenswrapper[4931]: I1125 21:51:34.989409 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:34Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.002648 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.011980 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.031609 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"message\\\":\\\"1125 21:51:33.767213 6140 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:33.767153 6140 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 21:51:33.767233 6140 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 21:51:33.767315 6140 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 21:51:33.767390 6140 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 21:51:33.767464 6140 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 21:51:33.767466 6140 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:33.767481 6140 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:33.767496 6140 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:33.767502 6140 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:33.767562 6140 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 21:51:33.767960 6140 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:33.768016 6140 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 21:51:33.768063 6140 factory.go:656] Stopping watch factory\\\\nI1125 21:51:33.768081 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1125 21:51:33.768105 6140 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.051010 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.059172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.059224 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.059241 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.059291 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.059308 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.069577 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.083333 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.096375 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.116962 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.142510 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.159422 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.161984 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.162046 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.162062 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.162086 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.162103 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.176124 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.196128 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.212342 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.231129 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.248633 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.267302 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.267349 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.267360 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.267376 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.267389 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.269107 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.284792 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.371096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.371142 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.371157 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.371176 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.371192 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.422141 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:35 crc kubenswrapper[4931]: E1125 21:51:35.422332 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.473950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.473989 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.474000 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.474018 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.474030 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.576908 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.576943 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.576951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.576964 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.576975 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.680522 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.680607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.680631 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.680662 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.680682 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.762659 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/0.log" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.768574 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.769049 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.775418 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" event={"ID":"53c90ad5-9f43-44c7-b122-0303c788f562","Type":"ContainerStarted","Data":"9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.782678 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.782716 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.782727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.782744 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.782757 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.794700 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.817813 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.838481 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.858715 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.876430 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.885866 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.886111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.886261 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.886425 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.886570 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.894156 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.911801 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.918389 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf"] Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.919750 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.925015 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.925065 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.932952 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.946100 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.959965 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.977791 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.989522 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.989575 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.989593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.989618 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.989636 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:35Z","lastTransitionTime":"2025-11-25T21:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:35 crc kubenswrapper[4931]: I1125 21:51:35.993858 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:35Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.006568 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.021572 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.044737 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/deeb760e-2656-4416-959f-11ba1746d6a2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.044804 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/deeb760e-2656-4416-959f-11ba1746d6a2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.044861 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdgk6\" (UniqueName: \"kubernetes.io/projected/deeb760e-2656-4416-959f-11ba1746d6a2-kube-api-access-vdgk6\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.044954 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/deeb760e-2656-4416-959f-11ba1746d6a2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.056904 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"message\\\":\\\"1125 21:51:33.767213 6140 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:33.767153 6140 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 21:51:33.767233 6140 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 21:51:33.767315 6140 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 21:51:33.767390 6140 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 21:51:33.767464 6140 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 21:51:33.767466 6140 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:33.767481 6140 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:33.767496 6140 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:33.767502 6140 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:33.767562 6140 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 21:51:33.767960 6140 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:33.768016 6140 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 21:51:33.768063 6140 factory.go:656] Stopping watch factory\\\\nI1125 21:51:33.768081 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1125 21:51:33.768105 6140 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.091383 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"message\\\":\\\"1125 21:51:33.767213 6140 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:33.767153 6140 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 21:51:33.767233 6140 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 21:51:33.767315 6140 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 21:51:33.767390 6140 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 21:51:33.767464 6140 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 21:51:33.767466 6140 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:33.767481 6140 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:33.767496 6140 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:33.767502 6140 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:33.767562 6140 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 21:51:33.767960 6140 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:33.768016 6140 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 21:51:33.768063 6140 factory.go:656] Stopping watch factory\\\\nI1125 21:51:33.768081 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1125 21:51:33.768105 6140 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.092404 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.092442 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.092456 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.092472 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.092484 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:36Z","lastTransitionTime":"2025-11-25T21:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.106539 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.123367 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.136921 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.146007 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdgk6\" (UniqueName: \"kubernetes.io/projected/deeb760e-2656-4416-959f-11ba1746d6a2-kube-api-access-vdgk6\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.146090 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/deeb760e-2656-4416-959f-11ba1746d6a2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.147001 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/deeb760e-2656-4416-959f-11ba1746d6a2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.147114 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/deeb760e-2656-4416-959f-11ba1746d6a2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.147140 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/deeb760e-2656-4416-959f-11ba1746d6a2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.147691 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/deeb760e-2656-4416-959f-11ba1746d6a2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.152171 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.161706 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/deeb760e-2656-4416-959f-11ba1746d6a2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.169065 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.178213 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdgk6\" (UniqueName: \"kubernetes.io/projected/deeb760e-2656-4416-959f-11ba1746d6a2-kube-api-access-vdgk6\") pod \"ovnkube-control-plane-749d76644c-4hqzf\" (UID: \"deeb760e-2656-4416-959f-11ba1746d6a2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.192531 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.195623 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.195688 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.195707 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.195733 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.195751 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:36Z","lastTransitionTime":"2025-11-25T21:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.208087 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.222388 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.244694 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.244808 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.265075 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: W1125 21:51:36.266470 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddeeb760e_2656_4416_959f_11ba1746d6a2.slice/crio-fa4aea5914e2e94e89f20acbdeb8cf302d6777781f8319d88b47609fb3eab213 WatchSource:0}: Error finding container fa4aea5914e2e94e89f20acbdeb8cf302d6777781f8319d88b47609fb3eab213: Status 404 returned error can't find the container with id fa4aea5914e2e94e89f20acbdeb8cf302d6777781f8319d88b47609fb3eab213 Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.288506 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.299053 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.299134 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.299159 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.299185 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.299204 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:36Z","lastTransitionTime":"2025-11-25T21:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.314345 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.331505 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.371989 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.401399 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.401455 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.401470 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.401493 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.401508 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:36Z","lastTransitionTime":"2025-11-25T21:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.405556 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.421920 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.421937 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:36 crc kubenswrapper[4931]: E1125 21:51:36.422054 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:36 crc kubenswrapper[4931]: E1125 21:51:36.422189 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.510953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.511026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.511045 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.511070 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.511089 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:36Z","lastTransitionTime":"2025-11-25T21:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.614519 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.614590 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.614688 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.614713 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.614733 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:36Z","lastTransitionTime":"2025-11-25T21:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.717101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.717143 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.717156 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.717171 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.717184 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:36Z","lastTransitionTime":"2025-11-25T21:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.781312 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/1.log" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.782304 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/0.log" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.786672 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564" exitCode=1 Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.786799 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.786915 4931 scope.go:117] "RemoveContainer" containerID="508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.787449 4931 scope.go:117] "RemoveContainer" containerID="c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564" Nov 25 21:51:36 crc kubenswrapper[4931]: E1125 21:51:36.787588 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.790071 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" event={"ID":"deeb760e-2656-4416-959f-11ba1746d6a2","Type":"ContainerStarted","Data":"18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.790142 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" event={"ID":"deeb760e-2656-4416-959f-11ba1746d6a2","Type":"ContainerStarted","Data":"77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.790173 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" event={"ID":"deeb760e-2656-4416-959f-11ba1746d6a2","Type":"ContainerStarted","Data":"fa4aea5914e2e94e89f20acbdeb8cf302d6777781f8319d88b47609fb3eab213"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.807594 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.819951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.820024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.820048 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.820077 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.820099 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:36Z","lastTransitionTime":"2025-11-25T21:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.823216 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.851924 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"message\\\":\\\"1125 21:51:33.767213 6140 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:33.767153 6140 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 21:51:33.767233 6140 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 21:51:33.767315 6140 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 21:51:33.767390 6140 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 21:51:33.767464 6140 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 21:51:33.767466 6140 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:33.767481 6140 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:33.767496 6140 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:33.767502 6140 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:33.767562 6140 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 21:51:33.767960 6140 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:33.768016 6140 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 21:51:33.768063 6140 factory.go:656] Stopping watch factory\\\\nI1125 21:51:33.768081 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1125 21:51:33.768105 6140 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.869829 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.883227 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.893277 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.906558 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.920659 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.922678 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.922719 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.922732 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.922749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.922760 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:36Z","lastTransitionTime":"2025-11-25T21:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.953949 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.967696 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.978448 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:36 crc kubenswrapper[4931]: I1125 21:51:36.996267 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:36Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.007314 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.017971 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.024661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.024693 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.024702 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.024714 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.024724 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.031904 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.045653 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.085462 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.126492 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.126976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.126998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.127006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.127022 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.127033 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.168123 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.209463 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.229811 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.229904 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.229923 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.229949 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.229966 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.247728 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.286643 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.324321 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.332132 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.332198 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.332219 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.332245 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.332263 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.380641 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"message\\\":\\\"1125 21:51:33.767213 6140 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:33.767153 6140 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 21:51:33.767233 6140 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 21:51:33.767315 6140 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 21:51:33.767390 6140 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 21:51:33.767464 6140 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 21:51:33.767466 6140 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:33.767481 6140 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:33.767496 6140 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:33.767502 6140 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:33.767562 6140 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 21:51:33.767960 6140 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:33.768016 6140 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 21:51:33.768063 6140 factory.go:656] Stopping watch factory\\\\nI1125 21:51:33.768081 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1125 21:51:33.768105 6140 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.412116 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.422076 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:37 crc kubenswrapper[4931]: E1125 21:51:37.422235 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.435672 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.435722 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.435739 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.435761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.435779 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.456365 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.492387 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.528501 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.540640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.540690 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.540711 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.540740 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.540765 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.574084 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.622611 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.644337 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.644406 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.644425 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.644451 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.644471 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.653110 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.690147 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:37Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.748231 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.748296 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.748317 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.748348 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.748375 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.796568 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/1.log" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.851529 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.851586 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.851604 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.851627 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.851645 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.955091 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.955187 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.955215 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.955248 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:37 crc kubenswrapper[4931]: I1125 21:51:37.955273 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:37Z","lastTransitionTime":"2025-11-25T21:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.058783 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.058869 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.058931 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.058963 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.058985 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.162476 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.162537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.162563 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.162591 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.162632 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.184198 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.184331 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.184392 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.184540 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.184663 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:54.184640296 +0000 UTC m=+52.541152455 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.185267 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.185354 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:51:54.185294334 +0000 UTC m=+52.541806503 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.185553 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:54.18551347 +0000 UTC m=+52.542025629 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.210422 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7v4p9"] Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.211278 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.211375 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.228864 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.244579 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.265513 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.265580 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.265597 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.265623 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.265641 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.272640 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.285572 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.285670 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.285746 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.285772 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.285786 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.285843 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:54.285827683 +0000 UTC m=+52.642339812 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.285861 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.285930 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.285952 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.286006 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 21:51:54.285988858 +0000 UTC m=+52.642501027 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.291961 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.315709 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.337676 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.357682 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.369789 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.369828 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.369842 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.369862 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.369905 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.386734 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.386593 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.386789 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2n5l\" (UniqueName: \"kubernetes.io/projected/779e1587-ca2f-4330-91dd-7251714f63bc-kube-api-access-w2n5l\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.405598 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.421624 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.421641 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.421818 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.422108 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.422303 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.440420 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.454801 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.472096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.472134 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.472148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.472168 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.472182 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.473635 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.488497 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.488549 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2n5l\" (UniqueName: \"kubernetes.io/projected/779e1587-ca2f-4330-91dd-7251714f63bc-kube-api-access-w2n5l\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.488940 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.488992 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs podName:779e1587-ca2f-4330-91dd-7251714f63bc nodeName:}" failed. No retries permitted until 2025-11-25 21:51:38.988975007 +0000 UTC m=+37.345487136 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs") pod "network-metrics-daemon-7v4p9" (UID: "779e1587-ca2f-4330-91dd-7251714f63bc") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.498299 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.533049 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2n5l\" (UniqueName: \"kubernetes.io/projected/779e1587-ca2f-4330-91dd-7251714f63bc-kube-api-access-w2n5l\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.536678 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.545208 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.545239 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.545251 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.545269 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.545281 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.555968 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"message\\\":\\\"1125 21:51:33.767213 6140 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:33.767153 6140 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 21:51:33.767233 6140 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 21:51:33.767315 6140 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 21:51:33.767390 6140 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 21:51:33.767464 6140 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 21:51:33.767466 6140 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:33.767481 6140 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:33.767496 6140 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:33.767502 6140 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:33.767562 6140 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 21:51:33.767960 6140 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:33.768016 6140 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 21:51:33.768063 6140 factory.go:656] Stopping watch factory\\\\nI1125 21:51:33.768081 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1125 21:51:33.768105 6140 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.561275 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.564501 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.564529 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.564538 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.564551 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.564560 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.573263 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.577111 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.580406 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.580445 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.580457 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.580474 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.580486 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.590831 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.595215 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.598350 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.598381 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.598395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.598410 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.598422 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.603592 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.609450 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.612669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.612723 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.612741 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.612765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.612782 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.614913 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.625068 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.625174 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.626279 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.626763 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.626790 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.626801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.626816 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.626827 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.638612 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.648710 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.659298 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.670295 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.704991 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.728804 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.728868 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.728923 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.728953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.728975 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.751735 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.790803 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.827918 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.831928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.831982 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.832001 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.832024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.832041 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.873087 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.921847 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508b670c25a7a9ca9b374ab2c94aca1633debd9cc912389714b5c33ac040a352\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"message\\\":\\\"1125 21:51:33.767213 6140 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:33.767153 6140 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 21:51:33.767233 6140 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 21:51:33.767315 6140 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 21:51:33.767390 6140 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 21:51:33.767464 6140 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 21:51:33.767466 6140 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:33.767481 6140 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:33.767496 6140 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:33.767502 6140 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:33.767562 6140 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 21:51:33.767960 6140 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:33.768016 6140 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 21:51:33.768063 6140 factory.go:656] Stopping watch factory\\\\nI1125 21:51:33.768081 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1125 21:51:33.768105 6140 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.935608 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.935709 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.935731 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.935755 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.935831 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:38Z","lastTransitionTime":"2025-11-25T21:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.963107 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.992744 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:38Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:38 crc kubenswrapper[4931]: I1125 21:51:38.993117 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.993215 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:38 crc kubenswrapper[4931]: E1125 21:51:38.993310 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs podName:779e1587-ca2f-4330-91dd-7251714f63bc nodeName:}" failed. No retries permitted until 2025-11-25 21:51:39.993285344 +0000 UTC m=+38.349797493 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs") pod "network-metrics-daemon-7v4p9" (UID: "779e1587-ca2f-4330-91dd-7251714f63bc") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.029042 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:39Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.039068 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.039149 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.039176 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.039210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.039229 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.070697 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:39Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.142524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.142563 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.142576 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.142594 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.142605 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.246671 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.246731 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.246748 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.246779 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.246797 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.350528 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.350601 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.350625 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.350703 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.350730 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.421970 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.421990 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:39 crc kubenswrapper[4931]: E1125 21:51:39.422245 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:39 crc kubenswrapper[4931]: E1125 21:51:39.422367 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.453742 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.453815 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.453841 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.453983 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.454034 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.556992 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.557055 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.557077 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.557108 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.557132 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.660054 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.660130 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.660150 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.660174 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.660195 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.763902 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.763958 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.763977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.763999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.764016 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.867600 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.867674 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.867693 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.867717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.867733 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.970760 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.970817 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.970835 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.970859 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:39 crc kubenswrapper[4931]: I1125 21:51:39.970918 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:39Z","lastTransitionTime":"2025-11-25T21:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.004497 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:40 crc kubenswrapper[4931]: E1125 21:51:40.004684 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:40 crc kubenswrapper[4931]: E1125 21:51:40.004760 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs podName:779e1587-ca2f-4330-91dd-7251714f63bc nodeName:}" failed. No retries permitted until 2025-11-25 21:51:42.004737675 +0000 UTC m=+40.361249824 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs") pod "network-metrics-daemon-7v4p9" (UID: "779e1587-ca2f-4330-91dd-7251714f63bc") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.073756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.073824 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.073843 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.073869 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.073910 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:40Z","lastTransitionTime":"2025-11-25T21:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.177437 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.177511 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.177537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.177567 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.177590 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:40Z","lastTransitionTime":"2025-11-25T21:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.281165 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.281234 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.281264 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.281309 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.281334 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:40Z","lastTransitionTime":"2025-11-25T21:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.383486 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.383524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.383536 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.383553 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.383565 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:40Z","lastTransitionTime":"2025-11-25T21:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.421535 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.421596 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:40 crc kubenswrapper[4931]: E1125 21:51:40.421630 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:40 crc kubenswrapper[4931]: E1125 21:51:40.421724 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.486349 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.486408 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.486425 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.486449 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.486464 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:40Z","lastTransitionTime":"2025-11-25T21:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.589376 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.589441 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.589459 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.589483 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.589502 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:40Z","lastTransitionTime":"2025-11-25T21:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.691864 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.691946 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.691963 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.691987 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.692004 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:40Z","lastTransitionTime":"2025-11-25T21:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.795006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.795137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.795177 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.795243 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.795265 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:40Z","lastTransitionTime":"2025-11-25T21:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.879543 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.880803 4931 scope.go:117] "RemoveContainer" containerID="c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564" Nov 25 21:51:40 crc kubenswrapper[4931]: E1125 21:51:40.881251 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.898241 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.898317 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.898340 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.898370 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.898393 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:40Z","lastTransitionTime":"2025-11-25T21:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.903933 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:40Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.920659 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:40Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.937610 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:40Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.959397 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:40Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.981822 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:40Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:40 crc kubenswrapper[4931]: I1125 21:51:40.999705 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:40Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.002433 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.002466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.002479 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.002495 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.002508 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.025194 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.053247 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.075901 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.094477 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.104986 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.105042 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.105061 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.105087 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.105107 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.111761 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.148434 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.169393 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.188851 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.205716 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.208363 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.208407 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.208418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.208433 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.208445 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.222015 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.243599 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:41Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.312106 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.312167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.312178 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.312199 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.312211 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.415866 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.416006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.416030 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.416068 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.416088 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.422405 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.422422 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:41 crc kubenswrapper[4931]: E1125 21:51:41.422597 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:41 crc kubenswrapper[4931]: E1125 21:51:41.422761 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.519332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.519386 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.519407 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.519432 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.519449 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.622245 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.622291 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.622300 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.622314 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.622324 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.725562 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.725639 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.725660 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.725685 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.725704 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.828927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.828992 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.829011 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.829036 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.829055 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.932076 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.932381 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.932391 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.932405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:41 crc kubenswrapper[4931]: I1125 21:51:41.932416 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:41Z","lastTransitionTime":"2025-11-25T21:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.026988 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:42 crc kubenswrapper[4931]: E1125 21:51:42.027179 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:42 crc kubenswrapper[4931]: E1125 21:51:42.027263 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs podName:779e1587-ca2f-4330-91dd-7251714f63bc nodeName:}" failed. No retries permitted until 2025-11-25 21:51:46.027240548 +0000 UTC m=+44.383752697 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs") pod "network-metrics-daemon-7v4p9" (UID: "779e1587-ca2f-4330-91dd-7251714f63bc") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.035551 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.035623 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.035649 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.035680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.035701 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.138996 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.139052 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.139068 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.139089 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.139105 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.242152 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.242194 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.242204 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.242221 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.242235 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.345210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.345352 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.345387 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.345420 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.345443 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.422552 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.422573 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:42 crc kubenswrapper[4931]: E1125 21:51:42.422843 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:42 crc kubenswrapper[4931]: E1125 21:51:42.423430 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.442668 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.448027 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.448096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.448116 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.448196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.448216 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.474127 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.491126 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.508723 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.528851 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.546294 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.554382 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.554434 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.554455 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.554485 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.554509 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.583678 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.606679 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.625335 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.642940 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.657870 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.657982 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.658005 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.658031 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.658051 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.665634 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.688762 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.707776 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.726648 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.742903 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.761470 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.761910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.761960 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.761977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.762003 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.762022 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.780739 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:42Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.865293 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.865348 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.865366 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.865389 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.865406 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.968218 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.968274 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.968296 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.968331 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:42 crc kubenswrapper[4931]: I1125 21:51:42.968352 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:42Z","lastTransitionTime":"2025-11-25T21:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.071847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.071941 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.071963 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.071992 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.072013 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:43Z","lastTransitionTime":"2025-11-25T21:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.175779 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.175831 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.175847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.175872 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.175919 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:43Z","lastTransitionTime":"2025-11-25T21:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.278456 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.278502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.278518 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.278540 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.278557 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:43Z","lastTransitionTime":"2025-11-25T21:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.382324 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.382389 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.382408 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.382432 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.382450 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:43Z","lastTransitionTime":"2025-11-25T21:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.422198 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.422372 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:43 crc kubenswrapper[4931]: E1125 21:51:43.422588 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:43 crc kubenswrapper[4931]: E1125 21:51:43.422771 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.485603 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.485653 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.485670 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.485696 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.485714 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:43Z","lastTransitionTime":"2025-11-25T21:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.588991 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.589049 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.589072 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.589099 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.589119 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:43Z","lastTransitionTime":"2025-11-25T21:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.691764 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.691840 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.691863 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.691924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.691942 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:43Z","lastTransitionTime":"2025-11-25T21:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.794198 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.794264 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.794284 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.794316 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.794341 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:43Z","lastTransitionTime":"2025-11-25T21:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.897069 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.897129 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.897148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.897170 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:43 crc kubenswrapper[4931]: I1125 21:51:43.897186 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:43Z","lastTransitionTime":"2025-11-25T21:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:43.999944 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:43.999997 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.000014 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.000038 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.000056 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.103318 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.103385 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.103402 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.103432 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.103452 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.207405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.207441 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.207452 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.207468 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.207506 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.322734 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.322799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.322820 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.322847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.322865 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.422051 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.422134 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:44 crc kubenswrapper[4931]: E1125 21:51:44.422239 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:44 crc kubenswrapper[4931]: E1125 21:51:44.422357 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.426426 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.426631 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.426801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.427037 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.427214 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.531036 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.531111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.531139 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.531186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.531210 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.635491 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.636112 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.636318 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.636500 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.636651 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.740631 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.740710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.740727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.740752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.740774 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.844020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.844095 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.844115 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.844142 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.844161 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.946549 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.946598 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.946614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.946630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:44 crc kubenswrapper[4931]: I1125 21:51:44.946643 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:44Z","lastTransitionTime":"2025-11-25T21:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.050006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.050238 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.050363 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.050628 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.050765 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.154096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.154194 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.154219 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.154252 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.154279 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.257315 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.257377 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.257394 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.257422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.257442 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.360736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.360801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.360819 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.360843 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.360861 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.421935 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.422065 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:45 crc kubenswrapper[4931]: E1125 21:51:45.422139 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:45 crc kubenswrapper[4931]: E1125 21:51:45.422228 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.464106 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.464181 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.464206 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.464235 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.464266 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.568135 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.568201 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.568315 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.568679 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.568743 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.673680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.673779 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.673797 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.673853 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.673902 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.777014 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.777079 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.777102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.777128 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.777149 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.880600 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.880660 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.880682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.880710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.880732 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.983132 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.983198 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.983220 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.983249 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:45 crc kubenswrapper[4931]: I1125 21:51:45.983271 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:45Z","lastTransitionTime":"2025-11-25T21:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.074821 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:46 crc kubenswrapper[4931]: E1125 21:51:46.075071 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:46 crc kubenswrapper[4931]: E1125 21:51:46.075143 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs podName:779e1587-ca2f-4330-91dd-7251714f63bc nodeName:}" failed. No retries permitted until 2025-11-25 21:51:54.075117899 +0000 UTC m=+52.431630058 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs") pod "network-metrics-daemon-7v4p9" (UID: "779e1587-ca2f-4330-91dd-7251714f63bc") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.085822 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.085866 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.085955 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.085979 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.085997 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:46Z","lastTransitionTime":"2025-11-25T21:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.189494 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.189588 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.189614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.189643 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.189663 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:46Z","lastTransitionTime":"2025-11-25T21:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.291967 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.292018 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.292035 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.292059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.292078 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:46Z","lastTransitionTime":"2025-11-25T21:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.395130 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.395275 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.395306 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.395336 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.395358 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:46Z","lastTransitionTime":"2025-11-25T21:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.421834 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.421937 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:46 crc kubenswrapper[4931]: E1125 21:51:46.422147 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:46 crc kubenswrapper[4931]: E1125 21:51:46.422197 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.497549 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.497612 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.497631 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.497659 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.497683 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:46Z","lastTransitionTime":"2025-11-25T21:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.601107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.601161 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.601180 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.601204 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.601221 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:46Z","lastTransitionTime":"2025-11-25T21:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.703458 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.703518 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.703541 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.703568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.703590 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:46Z","lastTransitionTime":"2025-11-25T21:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.807054 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.807112 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.807136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.807169 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.807190 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:46Z","lastTransitionTime":"2025-11-25T21:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.909696 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.909759 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.909776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.909801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:46 crc kubenswrapper[4931]: I1125 21:51:46.909819 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:46Z","lastTransitionTime":"2025-11-25T21:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.013143 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.013199 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.013215 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.013240 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.013262 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.116662 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.116715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.116737 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.116768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.116787 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.219988 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.220050 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.220069 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.220094 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.220110 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.327111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.327185 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.327205 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.327227 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.327245 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.422466 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.422471 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:47 crc kubenswrapper[4931]: E1125 21:51:47.422634 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:47 crc kubenswrapper[4931]: E1125 21:51:47.422783 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.429703 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.429762 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.429780 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.429803 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.429820 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.532833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.532927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.532948 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.532973 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.532991 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.637661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.637743 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.637769 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.637796 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.637822 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.741112 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.741170 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.741187 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.741214 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.741233 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.845757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.845833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.845857 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.845915 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.845935 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.948860 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.949002 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.949026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.949054 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:47 crc kubenswrapper[4931]: I1125 21:51:47.949077 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:47Z","lastTransitionTime":"2025-11-25T21:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.052041 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.052109 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.052127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.052152 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.052170 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.155575 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.155644 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.155668 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.155697 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.155720 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.258691 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.258755 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.258773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.258798 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.258817 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.362542 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.362630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.362658 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.362718 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.362742 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.422220 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:48 crc kubenswrapper[4931]: E1125 21:51:48.422412 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.422446 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:48 crc kubenswrapper[4931]: E1125 21:51:48.422628 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.466311 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.466394 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.466429 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.466460 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.466481 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.570200 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.570285 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.570310 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.570406 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.570442 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.645103 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.645174 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.645193 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.645220 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.645246 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: E1125 21:51:48.665987 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:48Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.671464 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.671524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.671543 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.671567 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.671585 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: E1125 21:51:48.691250 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:48Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.696303 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.696386 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.696408 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.696440 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.696464 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: E1125 21:51:48.717283 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:48Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.722075 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.722137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.722158 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.722183 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.722201 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: E1125 21:51:48.742007 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:48Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.746484 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.746534 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.746553 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.746574 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.746591 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: E1125 21:51:48.764779 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:48Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:48 crc kubenswrapper[4931]: E1125 21:51:48.765013 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.766810 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.766850 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.766866 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.766922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.766940 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.869403 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.869468 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.869487 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.869515 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.869534 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.972766 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.972836 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.972848 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.972869 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:48 crc kubenswrapper[4931]: I1125 21:51:48.972904 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:48Z","lastTransitionTime":"2025-11-25T21:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.075905 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.075966 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.075983 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.076007 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.076025 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:49Z","lastTransitionTime":"2025-11-25T21:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.185520 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.185580 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.185601 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.185626 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.185644 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:49Z","lastTransitionTime":"2025-11-25T21:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.289668 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.289734 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.289751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.289779 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.289796 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:49Z","lastTransitionTime":"2025-11-25T21:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.392823 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.392907 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.392927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.392950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.392968 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:49Z","lastTransitionTime":"2025-11-25T21:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.421868 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.421868 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:49 crc kubenswrapper[4931]: E1125 21:51:49.422098 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:49 crc kubenswrapper[4931]: E1125 21:51:49.422219 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.496922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.496971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.496988 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.497010 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.497029 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:49Z","lastTransitionTime":"2025-11-25T21:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.600739 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.600800 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.600824 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.600853 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.600906 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:49Z","lastTransitionTime":"2025-11-25T21:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.703818 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.703971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.704034 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.704059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.704076 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:49Z","lastTransitionTime":"2025-11-25T21:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.807545 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.807611 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.807629 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.807656 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.807674 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:49Z","lastTransitionTime":"2025-11-25T21:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.851607 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.865763 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.877435 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:49Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.895026 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:49Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.911552 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:49Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.912032 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.912222 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.912262 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.912469 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.912569 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:49Z","lastTransitionTime":"2025-11-25T21:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.937812 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:49Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.969452 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:49Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:49 crc kubenswrapper[4931]: I1125 21:51:49.990004 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:49Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.012108 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.017661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.017742 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.017768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.017800 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.017824 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.032966 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.068216 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.090805 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.110193 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.121503 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.121587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.121607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.121637 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.121660 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.130250 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.148707 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.172504 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.195278 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.217709 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.224775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.224843 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.224870 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.224939 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.224966 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.238096 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:50Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.328064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.328126 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.328145 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.328169 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.328187 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.422139 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.422156 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:50 crc kubenswrapper[4931]: E1125 21:51:50.422404 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:50 crc kubenswrapper[4931]: E1125 21:51:50.422526 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.433583 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.433675 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.433697 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.433725 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.433749 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.537165 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.537221 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.537240 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.537266 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.537288 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.640607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.640975 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.641157 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.641336 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.641656 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.744915 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.744973 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.744991 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.745013 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.745029 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.848227 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.848600 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.848769 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.848951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.849110 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.952504 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.952590 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.952604 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.952624 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:50 crc kubenswrapper[4931]: I1125 21:51:50.952638 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:50Z","lastTransitionTime":"2025-11-25T21:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.055607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.055665 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.055680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.055713 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.055733 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.159050 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.159116 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.159139 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.159168 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.159193 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.261835 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.261921 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.261936 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.261952 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.261964 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.365322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.365379 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.365397 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.365422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.365442 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.422044 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.422123 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:51 crc kubenswrapper[4931]: E1125 21:51:51.422234 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:51 crc kubenswrapper[4931]: E1125 21:51:51.422370 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.469289 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.469351 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.469399 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.469425 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.469445 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.572906 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.572972 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.572995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.573023 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.573044 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.675186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.675248 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.675267 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.675289 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.675307 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.778264 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.778326 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.778344 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.778371 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.778390 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.881632 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.881705 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.881730 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.881767 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.881793 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.985708 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.985774 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.985793 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.985818 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:51 crc kubenswrapper[4931]: I1125 21:51:51.985835 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:51Z","lastTransitionTime":"2025-11-25T21:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.089428 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.089483 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.089504 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.089527 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.089546 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:52Z","lastTransitionTime":"2025-11-25T21:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.192071 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.192140 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.192163 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.192196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.192215 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:52Z","lastTransitionTime":"2025-11-25T21:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.295721 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.295778 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.295796 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.295820 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.295839 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:52Z","lastTransitionTime":"2025-11-25T21:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.399034 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.399090 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.399109 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.399133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.399151 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:52Z","lastTransitionTime":"2025-11-25T21:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.422949 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.423353 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:52 crc kubenswrapper[4931]: E1125 21:51:52.423654 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:52 crc kubenswrapper[4931]: E1125 21:51:52.423959 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.444284 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.479826 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.501119 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.503211 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.503247 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.503261 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.503279 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.503290 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:52Z","lastTransitionTime":"2025-11-25T21:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.515634 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.539039 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.555137 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.575067 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.594425 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.609659 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.609721 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.609745 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.609778 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.609802 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:52Z","lastTransitionTime":"2025-11-25T21:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.611732 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.631299 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.652568 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.667565 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.687474 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.703492 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.713067 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.713183 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.713211 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.713286 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.713312 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:52Z","lastTransitionTime":"2025-11-25T21:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.727597 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.760497 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.778975 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.797074 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:52Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.815759 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.815830 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.815850 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.815903 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.815921 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:52Z","lastTransitionTime":"2025-11-25T21:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.918091 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.918155 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.918174 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.918199 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:52 crc kubenswrapper[4931]: I1125 21:51:52.918229 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:52Z","lastTransitionTime":"2025-11-25T21:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.021312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.021393 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.021505 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.021588 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.021634 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.124790 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.124844 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.124861 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.124956 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.124975 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.228775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.228836 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.228852 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.228918 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.228937 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.332101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.332178 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.332196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.332227 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.332247 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.421768 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.421937 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:53 crc kubenswrapper[4931]: E1125 21:51:53.421948 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:53 crc kubenswrapper[4931]: E1125 21:51:53.422158 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.436293 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.436392 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.436415 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.436476 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.436494 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.539800 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.539865 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.539938 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.539970 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.539994 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.642971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.643032 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.643049 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.643071 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.643089 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.748125 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.748189 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.748206 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.748231 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.748249 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.850830 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.850955 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.850975 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.851002 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.851020 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.954106 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.954174 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.954193 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.954217 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:53 crc kubenswrapper[4931]: I1125 21:51:53.954235 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:53Z","lastTransitionTime":"2025-11-25T21:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.056868 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.056977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.056995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.057021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.057040 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.160646 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.160716 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.160733 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.160760 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.160786 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.165324 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.165563 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.165666 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs podName:779e1587-ca2f-4330-91dd-7251714f63bc nodeName:}" failed. No retries permitted until 2025-11-25 21:52:10.165640569 +0000 UTC m=+68.522152718 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs") pod "network-metrics-daemon-7v4p9" (UID: "779e1587-ca2f-4330-91dd-7251714f63bc") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.264042 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.264132 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.264156 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.264188 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.264210 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.266468 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.266620 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.266680 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.266773 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:52:26.266742163 +0000 UTC m=+84.623254312 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.266782 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.266802 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.266900 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:52:26.266847287 +0000 UTC m=+84.623359436 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.266928 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:52:26.266914838 +0000 UTC m=+84.623426987 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.367677 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.367737 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.367927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.367935 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.367966 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.367977 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.368016 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.368039 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.368121 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 21:52:26.368094844 +0000 UTC m=+84.724607003 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.367992 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.368149 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.368168 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.368182 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.368314 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.368375 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 21:52:26.368355121 +0000 UTC m=+84.724867280 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.422460 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.422524 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.422674 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:54 crc kubenswrapper[4931]: E1125 21:51:54.423313 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.423853 4931 scope.go:117] "RemoveContainer" containerID="c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.471561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.471826 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.471844 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.471869 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.471928 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.575213 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.575274 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.575294 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.575321 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.575340 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.678151 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.678197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.678210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.678230 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.678243 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.781520 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.781561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.781571 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.781587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.781598 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.874319 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/1.log" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.878779 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.879574 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.884316 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.884383 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.884413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.884442 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.884466 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.907623 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:54Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.926859 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:54Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.987071 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.987107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.987119 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.987144 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.987156 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:54Z","lastTransitionTime":"2025-11-25T21:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:54 crc kubenswrapper[4931]: I1125 21:51:54.993013 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:54Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.020365 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.035686 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.048406 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.063216 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.076912 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.089630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.089669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.089683 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.089704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.089720 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:55Z","lastTransitionTime":"2025-11-25T21:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.093485 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.119798 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.132209 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.149707 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.171626 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.189436 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.192583 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.192628 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.192645 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.192667 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.192681 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:55Z","lastTransitionTime":"2025-11-25T21:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.203566 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.217101 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.239421 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.257394 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.295267 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.295337 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.295357 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.295380 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.295394 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:55Z","lastTransitionTime":"2025-11-25T21:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.398020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.398062 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.398074 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.398091 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.398105 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:55Z","lastTransitionTime":"2025-11-25T21:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.422394 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.422417 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:55 crc kubenswrapper[4931]: E1125 21:51:55.422517 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:55 crc kubenswrapper[4931]: E1125 21:51:55.422624 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.500499 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.500532 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.500543 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.500558 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.500570 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:55Z","lastTransitionTime":"2025-11-25T21:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.603005 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.603064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.603080 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.603101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.603118 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:55Z","lastTransitionTime":"2025-11-25T21:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.705934 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.706020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.706043 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.706073 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.706095 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:55Z","lastTransitionTime":"2025-11-25T21:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.809006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.809043 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.809051 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.809064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.809073 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:55Z","lastTransitionTime":"2025-11-25T21:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.885364 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/2.log" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.886473 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/1.log" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.890709 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef" exitCode=1 Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.890764 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.890824 4931 scope.go:117] "RemoveContainer" containerID="c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.892435 4931 scope.go:117] "RemoveContainer" containerID="7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef" Nov 25 21:51:55 crc kubenswrapper[4931]: E1125 21:51:55.893020 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.912575 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.912629 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.912653 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.912677 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.912695 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:55Z","lastTransitionTime":"2025-11-25T21:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.931118 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.953100 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.971404 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:55 crc kubenswrapper[4931]: I1125 21:51:55.987152 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:55Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.009700 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.015369 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.015440 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.015459 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.015487 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.015505 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.032160 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.052852 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.072525 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.091476 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.109662 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.119062 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.119147 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.119166 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.119192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.119211 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.133351 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.151260 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.171352 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.193545 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.211404 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.221833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.221950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.221973 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.221998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.222017 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.225057 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.252840 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.282932 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c578aa8bd613c11c096e28634cc4d4c8287bf5c7f4f088ee8b0a844a433a3564\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"message\\\":\\\"actory.go:117\\\\nI1125 21:51:36.064092 6355 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 21:51:36.064142 6355 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 21:51:36.064167 6355 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:51:36.064371 6355 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 21:51:36.064738 6355 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 21:51:36.064790 6355 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 21:51:36.064825 6355 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:51:36.064852 6355 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 21:51:36.064906 6355 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 21:51:36.064868 6355 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:51:36.064936 6355 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 21:51:36.065009 6355 factory.go:656] Stopping watch factory\\\\nI1125 21:51:36.065026 6355 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:55Z\\\",\\\"message\\\":\\\"ccess event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1125 21:51:55.404069 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-7v4p9\\\\nI1125 21:51:55.404072 6589 services_controller.go:454] Service openshift-route-controller-manager/route-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1125 21:51:55.404080 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404086 6589 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404090 6589 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 21:51:55.404095 6589 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 21:51:55.404100 6589 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404108 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jjhc2\\\\nI1125 21:51:55.404112 6589 obj_retry.go:365] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.325339 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.325405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.325427 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.325455 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.325476 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.422056 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.422136 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:56 crc kubenswrapper[4931]: E1125 21:51:56.422272 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:56 crc kubenswrapper[4931]: E1125 21:51:56.422406 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.429709 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.429759 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.429788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.429813 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.429836 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.533018 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.533079 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.533096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.533122 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.533141 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.635854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.635953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.635973 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.635998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.636016 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.739112 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.739185 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.739206 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.739231 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.739248 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.842682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.842828 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.842854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.842932 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.842963 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.905755 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/2.log" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.911277 4931 scope.go:117] "RemoveContainer" containerID="7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef" Nov 25 21:51:56 crc kubenswrapper[4931]: E1125 21:51:56.911580 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.930394 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.946386 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.946452 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.946474 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.946503 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.946527 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:56Z","lastTransitionTime":"2025-11-25T21:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.954132 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.973983 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:56 crc kubenswrapper[4931]: I1125 21:51:56.992494 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:56Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.011633 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.031191 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.049925 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.049985 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.050003 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.050027 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.050044 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.051757 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.066589 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.080230 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.097499 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.128992 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:55Z\\\",\\\"message\\\":\\\"ccess event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1125 21:51:55.404069 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-7v4p9\\\\nI1125 21:51:55.404072 6589 services_controller.go:454] Service openshift-route-controller-manager/route-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1125 21:51:55.404080 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404086 6589 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404090 6589 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 21:51:55.404095 6589 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 21:51:55.404100 6589 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404108 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jjhc2\\\\nI1125 21:51:55.404112 6589 obj_retry.go:365] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.141935 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.152317 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.152371 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.152384 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.152405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.152422 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.159313 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.172166 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.202027 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.219562 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.234283 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.247308 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:57Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.255246 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.255274 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.255282 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.255297 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.255307 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.357629 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.357663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.357671 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.357684 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.357693 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.422439 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.422506 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:57 crc kubenswrapper[4931]: E1125 21:51:57.422575 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:57 crc kubenswrapper[4931]: E1125 21:51:57.422718 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.460814 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.460917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.460941 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.460969 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.460987 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.563736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.563789 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.563807 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.563829 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.563845 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.667825 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.667990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.668085 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.668123 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.668147 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.771547 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.771622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.771643 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.771669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.771687 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.874113 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.874209 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.874234 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.874264 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.874283 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.977573 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.977626 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.977647 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.977668 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:57 crc kubenswrapper[4931]: I1125 21:51:57.977684 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:57Z","lastTransitionTime":"2025-11-25T21:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.080927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.080997 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.081015 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.081041 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.081056 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.183727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.183794 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.183810 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.183835 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.183852 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.288183 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.288264 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.288288 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.288319 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.288341 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.391547 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.391641 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.391683 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.391719 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.391740 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.421915 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.421947 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:51:58 crc kubenswrapper[4931]: E1125 21:51:58.422123 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:51:58 crc kubenswrapper[4931]: E1125 21:51:58.422248 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.494813 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.494929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.494969 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.495000 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.495026 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.597544 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.597612 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.597633 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.597665 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.597692 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.700531 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.700599 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.700619 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.700650 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.700701 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.803782 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.803847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.803866 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.803935 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.803956 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.809661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.809722 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.809745 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.809773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.809795 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: E1125 21:51:58.831674 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:58Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.836811 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.836868 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.836890 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.836949 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.836975 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: E1125 21:51:58.856442 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:58Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.860803 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.860856 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.860878 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.860926 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.860943 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: E1125 21:51:58.883811 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:58Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.890191 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.890251 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.890273 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.890302 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.890324 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: E1125 21:51:58.914660 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:58Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.920568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.920628 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.920646 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.920669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.920687 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:58 crc kubenswrapper[4931]: E1125 21:51:58.943675 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:51:58Z is after 2025-08-24T17:21:41Z" Nov 25 21:51:58 crc kubenswrapper[4931]: E1125 21:51:58.943942 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.950995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.951104 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.951127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.951213 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:58 crc kubenswrapper[4931]: I1125 21:51:58.951276 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:58Z","lastTransitionTime":"2025-11-25T21:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.054482 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.054543 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.054561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.054584 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.054603 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.157717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.157990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.158009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.158038 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.158055 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.260630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.260675 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.260691 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.260715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.260733 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.363835 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.363968 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.363989 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.364021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.364038 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.421748 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.421842 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:51:59 crc kubenswrapper[4931]: E1125 21:51:59.422018 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:51:59 crc kubenswrapper[4931]: E1125 21:51:59.422132 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.467214 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.467301 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.467361 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.467393 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.467415 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.570395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.570509 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.570532 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.570561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.570583 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.673718 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.673777 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.673799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.673823 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.673841 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.777537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.777614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.777633 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.777659 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.777677 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.880833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.880926 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.880945 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.880968 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.880984 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.983818 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.983930 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.983958 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.983980 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:51:59 crc kubenswrapper[4931]: I1125 21:51:59.984027 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:51:59Z","lastTransitionTime":"2025-11-25T21:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.087498 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.087559 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.087579 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.087606 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.087628 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:00Z","lastTransitionTime":"2025-11-25T21:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.190862 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.190960 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.190978 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.191010 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.191029 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:00Z","lastTransitionTime":"2025-11-25T21:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.294354 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.294412 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.294431 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.294455 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.294473 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:00Z","lastTransitionTime":"2025-11-25T21:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.397175 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.397253 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.397294 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.397329 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.397354 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:00Z","lastTransitionTime":"2025-11-25T21:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.422213 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.422269 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:00 crc kubenswrapper[4931]: E1125 21:52:00.422375 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:00 crc kubenswrapper[4931]: E1125 21:52:00.422475 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.500408 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.500479 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.500506 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.500538 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.500560 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:00Z","lastTransitionTime":"2025-11-25T21:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.603418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.603495 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.603518 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.603546 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.603567 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:00Z","lastTransitionTime":"2025-11-25T21:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.706239 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.706286 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.706305 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.706327 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.706345 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:00Z","lastTransitionTime":"2025-11-25T21:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.809431 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.809485 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.809503 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.809526 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.809542 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:00Z","lastTransitionTime":"2025-11-25T21:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.912369 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.912434 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.912452 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.912476 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:00 crc kubenswrapper[4931]: I1125 21:52:00.912510 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:00Z","lastTransitionTime":"2025-11-25T21:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.016402 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.016507 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.016661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.016697 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.016768 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.119784 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.119917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.119950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.119979 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.120001 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.222312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.222361 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.222379 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.222403 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.222421 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.326163 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.326227 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.326250 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.326280 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.326302 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.421671 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.421698 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:01 crc kubenswrapper[4931]: E1125 21:52:01.421869 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:01 crc kubenswrapper[4931]: E1125 21:52:01.422035 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.430155 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.430202 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.430220 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.430242 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.430259 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.532951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.533080 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.533102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.533127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.533148 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.637039 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.637107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.637125 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.637151 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.637169 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.739810 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.739884 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.739974 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.740558 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.740615 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.844606 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.844706 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.844726 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.844751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.844770 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.947966 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.948042 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.948061 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.948085 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:01 crc kubenswrapper[4931]: I1125 21:52:01.948102 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:01Z","lastTransitionTime":"2025-11-25T21:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.050544 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.050590 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.050609 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.050631 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.050647 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.153087 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.153142 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.153154 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.153174 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.153186 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.256316 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.256377 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.256395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.256420 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.256437 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.358685 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.358746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.358762 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.358787 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.358805 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.421642 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:02 crc kubenswrapper[4931]: E1125 21:52:02.421841 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.421649 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:02 crc kubenswrapper[4931]: E1125 21:52:02.422385 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.449519 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.462438 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.462492 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.462509 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.462536 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.462555 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.472798 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.492153 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.509514 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.533931 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.565172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.565236 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.565255 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.565279 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.565298 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.569852 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:55Z\\\",\\\"message\\\":\\\"ccess event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1125 21:51:55.404069 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-7v4p9\\\\nI1125 21:51:55.404072 6589 services_controller.go:454] Service openshift-route-controller-manager/route-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1125 21:51:55.404080 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404086 6589 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404090 6589 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 21:51:55.404095 6589 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 21:51:55.404100 6589 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404108 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jjhc2\\\\nI1125 21:51:55.404112 6589 obj_retry.go:365] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.603937 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.623034 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.637141 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.650919 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.670805 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.670841 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.670854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.670871 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.670908 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.670827 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.687746 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.708386 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.727593 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.745943 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.763253 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.773403 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.773445 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.773460 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.773478 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.773492 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.784541 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.801597 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:02Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.877003 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.877049 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.877063 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.877082 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.877095 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.979644 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.979951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.979966 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.979985 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:02 crc kubenswrapper[4931]: I1125 21:52:02.979998 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:02Z","lastTransitionTime":"2025-11-25T21:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.082751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.082802 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.082818 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.082841 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.082858 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:03Z","lastTransitionTime":"2025-11-25T21:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.185851 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.185950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.185971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.185999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.186016 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:03Z","lastTransitionTime":"2025-11-25T21:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.290140 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.290195 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.290212 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.290234 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.290251 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:03Z","lastTransitionTime":"2025-11-25T21:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.393383 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.393740 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.393946 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.394095 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.394223 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:03Z","lastTransitionTime":"2025-11-25T21:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.422041 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.422069 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:03 crc kubenswrapper[4931]: E1125 21:52:03.422235 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:03 crc kubenswrapper[4931]: E1125 21:52:03.422385 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.497662 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.497725 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.497744 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.497769 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.497786 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:03Z","lastTransitionTime":"2025-11-25T21:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.601207 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.601284 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.601312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.601343 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.601372 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:03Z","lastTransitionTime":"2025-11-25T21:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.704794 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.704879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.704940 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.704971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.705001 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:03Z","lastTransitionTime":"2025-11-25T21:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.808116 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.808181 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.808199 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.808223 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.808240 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:03Z","lastTransitionTime":"2025-11-25T21:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.910947 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.911005 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.911022 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.911046 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:03 crc kubenswrapper[4931]: I1125 21:52:03.911084 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:03Z","lastTransitionTime":"2025-11-25T21:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.013602 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.013669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.013686 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.013712 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.013731 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.117111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.117163 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.117177 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.117195 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.117208 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.219614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.219688 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.219718 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.219779 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.219800 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.323360 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.323677 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.323867 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.324125 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.324315 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.422140 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.422138 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:04 crc kubenswrapper[4931]: E1125 21:52:04.422360 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:04 crc kubenswrapper[4931]: E1125 21:52:04.422469 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.428532 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.428592 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.428614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.428640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.428656 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.531871 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.531964 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.531982 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.532007 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.532027 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.635120 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.635197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.635223 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.635255 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.635283 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.737847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.737950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.737970 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.737994 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.738012 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.840304 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.840368 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.840395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.840427 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.840448 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.942593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.942642 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.942661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.942694 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:04 crc kubenswrapper[4931]: I1125 21:52:04.942713 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:04Z","lastTransitionTime":"2025-11-25T21:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.046073 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.046146 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.046167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.046191 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.046208 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.149208 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.149619 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.149781 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.149969 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.150138 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.253698 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.253755 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.253772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.253795 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.253814 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.356662 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.356752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.356767 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.356787 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.356803 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.422473 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.422512 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:05 crc kubenswrapper[4931]: E1125 21:52:05.423642 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:05 crc kubenswrapper[4931]: E1125 21:52:05.423510 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.460255 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.460579 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.460750 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.460924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.461139 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.563637 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.563712 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.563737 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.563766 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.563787 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.667487 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.667617 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.667643 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.667675 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.667698 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.772840 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.772912 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.772924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.772941 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.772954 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.875833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.875869 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.875903 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.875922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.875937 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.978844 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.978919 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.978956 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.978976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:05 crc kubenswrapper[4931]: I1125 21:52:05.978988 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:05Z","lastTransitionTime":"2025-11-25T21:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.081022 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.081081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.081100 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.081124 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.081142 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:06Z","lastTransitionTime":"2025-11-25T21:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.184035 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.184089 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.184101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.184117 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.184131 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:06Z","lastTransitionTime":"2025-11-25T21:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.286833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.286895 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.286911 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.286961 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.286974 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:06Z","lastTransitionTime":"2025-11-25T21:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.390719 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.390782 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.390805 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.390833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.390855 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:06Z","lastTransitionTime":"2025-11-25T21:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.422004 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:06 crc kubenswrapper[4931]: E1125 21:52:06.422293 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.422033 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:06 crc kubenswrapper[4931]: E1125 21:52:06.422601 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.494507 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.494568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.494585 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.494608 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.494626 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:06Z","lastTransitionTime":"2025-11-25T21:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.597935 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.598019 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.598040 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.598066 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.598084 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:06Z","lastTransitionTime":"2025-11-25T21:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.701151 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.701225 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.701248 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.701317 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.701342 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:06Z","lastTransitionTime":"2025-11-25T21:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.804707 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.804762 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.804782 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.804807 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.804828 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:06Z","lastTransitionTime":"2025-11-25T21:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.907494 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.907559 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.907571 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.907586 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:06 crc kubenswrapper[4931]: I1125 21:52:06.907598 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:06Z","lastTransitionTime":"2025-11-25T21:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.010335 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.010395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.010414 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.010439 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.010455 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.120039 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.120102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.120121 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.120162 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.120185 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.222541 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.222598 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.222616 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.222638 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.222654 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.324794 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.324917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.324936 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.324963 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.324984 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.421930 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.422030 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:07 crc kubenswrapper[4931]: E1125 21:52:07.422065 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:07 crc kubenswrapper[4931]: E1125 21:52:07.422207 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.428350 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.428418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.428439 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.428462 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.428479 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.530703 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.530769 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.530789 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.530818 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.530837 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.635744 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.635781 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.635790 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.635807 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.635819 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.738598 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.738641 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.738649 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.738663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.738674 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.841356 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.841396 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.841408 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.841426 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.841438 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.947294 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.947401 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.947421 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.947483 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:07 crc kubenswrapper[4931]: I1125 21:52:07.947507 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:07Z","lastTransitionTime":"2025-11-25T21:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.050709 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.050755 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.050764 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.050799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.050809 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.153934 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.153977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.153990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.154004 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.154016 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.256368 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.256428 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.256441 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.256461 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.256473 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.359367 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.359424 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.359437 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.359457 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.359468 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.421972 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.421981 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:08 crc kubenswrapper[4931]: E1125 21:52:08.422119 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:08 crc kubenswrapper[4931]: E1125 21:52:08.422207 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.461941 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.461994 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.462006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.462026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.462038 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.564684 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.564764 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.564775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.564797 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.564810 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.668700 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.668761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.668772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.668797 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.668813 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.771668 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.771708 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.771720 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.771736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.771746 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.874508 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.874598 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.874616 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.874649 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.874673 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.977087 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.977138 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.977149 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.977168 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:08 crc kubenswrapper[4931]: I1125 21:52:08.977180 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:08Z","lastTransitionTime":"2025-11-25T21:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.080049 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.080102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.080114 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.080132 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.080147 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.182303 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.182363 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.182383 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.182408 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.182426 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.284466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.284500 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.284508 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.284520 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.284531 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.321332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.321370 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.321380 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.321394 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.321404 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: E1125 21:52:09.338655 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:09Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.343981 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.344042 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.344060 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.344085 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.344103 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: E1125 21:52:09.360131 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:09Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.364254 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.364292 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.364301 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.364316 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.364328 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: E1125 21:52:09.380522 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:09Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.385088 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.385122 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.385132 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.385148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.385161 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: E1125 21:52:09.404632 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:09Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.408621 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.408661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.408670 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.408685 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.408694 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.421970 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.421978 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:09 crc kubenswrapper[4931]: E1125 21:52:09.422180 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:09 crc kubenswrapper[4931]: E1125 21:52:09.422247 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:09 crc kubenswrapper[4931]: E1125 21:52:09.423107 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:09Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:09 crc kubenswrapper[4931]: E1125 21:52:09.423323 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.425284 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.425348 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.425372 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.425400 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.425423 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.527976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.528045 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.528070 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.528101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.528124 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.630702 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.630742 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.630752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.630768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.630797 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.733409 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.733464 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.733481 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.733504 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.733521 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.836475 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.836537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.836556 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.836581 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.836598 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.939297 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.939356 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.939372 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.939399 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:09 crc kubenswrapper[4931]: I1125 21:52:09.939416 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:09Z","lastTransitionTime":"2025-11-25T21:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.043090 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.043145 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.043168 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.043197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.043215 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.146095 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.146175 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.146192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.146215 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.146231 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.249083 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.249138 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.249162 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.249188 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.249208 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.256578 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:10 crc kubenswrapper[4931]: E1125 21:52:10.256738 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:52:10 crc kubenswrapper[4931]: E1125 21:52:10.256802 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs podName:779e1587-ca2f-4330-91dd-7251714f63bc nodeName:}" failed. No retries permitted until 2025-11-25 21:52:42.256780645 +0000 UTC m=+100.613292794 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs") pod "network-metrics-daemon-7v4p9" (UID: "779e1587-ca2f-4330-91dd-7251714f63bc") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.351664 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.351717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.351734 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.351762 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.351781 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.422253 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.422300 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:10 crc kubenswrapper[4931]: E1125 21:52:10.422437 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:10 crc kubenswrapper[4931]: E1125 21:52:10.422618 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.454443 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.454481 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.454498 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.454522 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.454539 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.557848 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.557917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.557929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.557945 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.557959 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.660728 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.660758 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.660765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.660778 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.660786 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.763002 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.763111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.763131 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.763160 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.763180 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.866654 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.866716 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.866732 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.866758 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.866775 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.969347 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.969385 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.969395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.969410 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:10 crc kubenswrapper[4931]: I1125 21:52:10.969421 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:10Z","lastTransitionTime":"2025-11-25T21:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.073643 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.073692 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.073710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.073731 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.073747 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:11Z","lastTransitionTime":"2025-11-25T21:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.176562 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.176633 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.176652 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.176676 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.176700 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:11Z","lastTransitionTime":"2025-11-25T21:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.280595 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.280660 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.280682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.280713 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.280736 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:11Z","lastTransitionTime":"2025-11-25T21:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.384290 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.384357 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.384379 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.384414 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.384481 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:11Z","lastTransitionTime":"2025-11-25T21:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.422572 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.422578 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:11 crc kubenswrapper[4931]: E1125 21:52:11.422787 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:11 crc kubenswrapper[4931]: E1125 21:52:11.423037 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.487983 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.488061 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.488078 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.488104 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.488123 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:11Z","lastTransitionTime":"2025-11-25T21:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.590242 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.590322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.590345 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.590386 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.590403 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:11Z","lastTransitionTime":"2025-11-25T21:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.693699 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.693764 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.693777 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.693793 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.693804 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:11Z","lastTransitionTime":"2025-11-25T21:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.796715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.796775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.796789 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.796810 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.796820 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:11Z","lastTransitionTime":"2025-11-25T21:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.899770 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.899911 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.899984 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.900009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:11 crc kubenswrapper[4931]: I1125 21:52:11.900032 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:11Z","lastTransitionTime":"2025-11-25T21:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.003064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.003102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.003110 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.003124 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.003134 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.105390 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.105418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.105426 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.105439 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.105448 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.208273 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.208310 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.208318 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.208352 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.208362 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.311237 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.311303 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.311326 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.311357 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.311380 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.414431 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.414493 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.414509 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.414535 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.414553 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.421712 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:12 crc kubenswrapper[4931]: E1125 21:52:12.421853 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.422127 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:12 crc kubenswrapper[4931]: E1125 21:52:12.422488 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.423102 4931 scope.go:117] "RemoveContainer" containerID="7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef" Nov 25 21:52:12 crc kubenswrapper[4931]: E1125 21:52:12.423541 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.434320 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.455723 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.472103 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.489834 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.508838 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:55Z\\\",\\\"message\\\":\\\"ccess event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1125 21:51:55.404069 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-7v4p9\\\\nI1125 21:51:55.404072 6589 services_controller.go:454] Service openshift-route-controller-manager/route-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1125 21:51:55.404080 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404086 6589 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404090 6589 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 21:51:55.404095 6589 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 21:51:55.404100 6589 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404108 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jjhc2\\\\nI1125 21:51:55.404112 6589 obj_retry.go:365] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.517131 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.517378 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.517561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.517751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.517936 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.522782 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.537324 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.550479 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.571476 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.588150 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.600899 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.617106 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.620417 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.620472 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.620485 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.620502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.620513 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.630164 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.643318 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.656066 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.668597 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.679063 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.692137 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:12Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.722898 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.722938 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.722948 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.722962 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.722971 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.825047 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.825330 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.825422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.825504 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.825574 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.927956 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.928036 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.928055 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.928084 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:12 crc kubenswrapper[4931]: I1125 21:52:12.928101 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:12Z","lastTransitionTime":"2025-11-25T21:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.031125 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.031163 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.031172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.031186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.031195 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.134799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.134858 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.134901 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.134924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.134941 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.237311 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.237405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.237431 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.237458 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.237481 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.340691 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.340752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.340765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.340783 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.340796 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.421564 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:13 crc kubenswrapper[4931]: E1125 21:52:13.421683 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.421585 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:13 crc kubenswrapper[4931]: E1125 21:52:13.421774 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.443912 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.443949 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.443961 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.443974 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.443984 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.546219 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.546282 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.546307 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.546330 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.546349 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.648665 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.648724 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.648738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.648757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.648769 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.751593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.751690 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.751710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.751734 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.751754 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.853815 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.853846 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.853854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.853867 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.853892 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.957311 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.957379 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.957396 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.957423 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:13 crc kubenswrapper[4931]: I1125 21:52:13.957442 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:13Z","lastTransitionTime":"2025-11-25T21:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.060019 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.060087 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.060114 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.060144 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.060168 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.162908 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.162978 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.163000 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.163028 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.163053 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.266361 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.266422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.266438 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.266463 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.266480 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.369410 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.369461 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.369478 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.369502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.369520 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.422582 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.422649 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:14 crc kubenswrapper[4931]: E1125 21:52:14.422728 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:14 crc kubenswrapper[4931]: E1125 21:52:14.422783 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.471757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.471801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.471813 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.471829 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.471842 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.574814 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.574861 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.574888 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.574922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.574938 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.677998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.678040 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.678054 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.678074 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.678087 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.781004 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.781041 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.781054 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.781070 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.781082 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.886568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.886609 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.886622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.886640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.886654 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.978290 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/0.log" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.978388 4931 generic.go:334] "Generic (PLEG): container finished" podID="ef281bd7-e1b1-41cd-b0a2-fa026829cf8f" containerID="fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4" exitCode=1 Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.978444 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xltbl" event={"ID":"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f","Type":"ContainerDied","Data":"fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.979176 4931 scope.go:117] "RemoveContainer" containerID="fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.988576 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.988793 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.988813 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.988828 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.988843 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:14Z","lastTransitionTime":"2025-11-25T21:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:14 crc kubenswrapper[4931]: I1125 21:52:14.992727 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:14Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.008376 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.018529 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.032002 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.045805 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.058910 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.070253 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.085280 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.090612 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.090639 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.090649 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.090663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.090673 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:15Z","lastTransitionTime":"2025-11-25T21:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.104020 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:55Z\\\",\\\"message\\\":\\\"ccess event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1125 21:51:55.404069 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-7v4p9\\\\nI1125 21:51:55.404072 6589 services_controller.go:454] Service openshift-route-controller-manager/route-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1125 21:51:55.404080 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404086 6589 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404090 6589 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 21:51:55.404095 6589 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 21:51:55.404100 6589 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404108 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jjhc2\\\\nI1125 21:51:55.404112 6589 obj_retry.go:365] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.125749 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.136431 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:52:14Z\\\",\\\"message\\\":\\\"2025-11-25T21:51:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449\\\\n2025-11-25T21:51:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449 to /host/opt/cni/bin/\\\\n2025-11-25T21:51:29Z [verbose] multus-daemon started\\\\n2025-11-25T21:51:29Z [verbose] Readiness Indicator file check\\\\n2025-11-25T21:52:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.146589 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.154827 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.168016 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.181344 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.192821 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.192849 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.192860 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.192891 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.192901 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:15Z","lastTransitionTime":"2025-11-25T21:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.193099 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.204651 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.215549 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:15Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.297497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.297557 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.297574 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.297597 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.297619 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:15Z","lastTransitionTime":"2025-11-25T21:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.405143 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.405208 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.405226 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.405253 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.405271 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:15Z","lastTransitionTime":"2025-11-25T21:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.422572 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:15 crc kubenswrapper[4931]: E1125 21:52:15.422747 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.423028 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:15 crc kubenswrapper[4931]: E1125 21:52:15.423136 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.507928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.507992 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.508010 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.508034 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.508051 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:15Z","lastTransitionTime":"2025-11-25T21:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.611414 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.611468 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.611486 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.611509 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.611526 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:15Z","lastTransitionTime":"2025-11-25T21:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.715608 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.715714 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.715737 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.715764 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.715785 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:15Z","lastTransitionTime":"2025-11-25T21:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.818213 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.818288 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.818312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.818342 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.818365 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:15Z","lastTransitionTime":"2025-11-25T21:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.921411 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.921470 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.921488 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.921512 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.921568 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:15Z","lastTransitionTime":"2025-11-25T21:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.983858 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/0.log" Nov 25 21:52:15 crc kubenswrapper[4931]: I1125 21:52:15.983962 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xltbl" event={"ID":"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f","Type":"ContainerStarted","Data":"0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.008014 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.023281 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.023328 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.023344 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.023367 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.023411 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.027709 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.043673 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.055572 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.069918 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.084792 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.100263 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.111913 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.125766 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.125814 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.125826 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.125843 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.125854 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.128944 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.140408 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.155010 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.164810 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.177318 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.203277 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:55Z\\\",\\\"message\\\":\\\"ccess event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1125 21:51:55.404069 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-7v4p9\\\\nI1125 21:51:55.404072 6589 services_controller.go:454] Service openshift-route-controller-manager/route-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1125 21:51:55.404080 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404086 6589 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404090 6589 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 21:51:55.404095 6589 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 21:51:55.404100 6589 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404108 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jjhc2\\\\nI1125 21:51:55.404112 6589 obj_retry.go:365] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.227524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.227788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.227900 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.228000 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.228079 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.231413 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.248952 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:52:14Z\\\",\\\"message\\\":\\\"2025-11-25T21:51:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449\\\\n2025-11-25T21:51:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449 to /host/opt/cni/bin/\\\\n2025-11-25T21:51:29Z [verbose] multus-daemon started\\\\n2025-11-25T21:51:29Z [verbose] Readiness Indicator file check\\\\n2025-11-25T21:52:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:52:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.260442 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.269859 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:16Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.330809 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.330839 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.330848 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.330903 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.330921 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.421936 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:16 crc kubenswrapper[4931]: E1125 21:52:16.422317 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.422040 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:16 crc kubenswrapper[4931]: E1125 21:52:16.422638 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.433419 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.436343 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.436378 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.436392 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.436409 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.538582 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.538622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.538631 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.538645 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.538654 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.640753 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.640786 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.640794 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.640807 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.640817 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.743113 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.743143 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.743152 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.743165 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.743174 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.844943 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.844967 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.844976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.844987 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.844996 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.946967 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.946998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.947008 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.947021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:16 crc kubenswrapper[4931]: I1125 21:52:16.947032 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:16Z","lastTransitionTime":"2025-11-25T21:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.049106 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.049150 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.049160 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.049177 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.049189 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.151706 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.151746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.151756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.151770 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.151782 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.254467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.254532 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.254550 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.254575 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.254592 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.357535 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.357598 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.357621 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.357648 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.357669 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.422502 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.422580 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:17 crc kubenswrapper[4931]: E1125 21:52:17.422741 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:17 crc kubenswrapper[4931]: E1125 21:52:17.422938 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.461022 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.461081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.461094 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.461112 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.461125 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.564315 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.564373 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.564383 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.564396 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.564423 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.667344 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.667400 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.667419 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.667446 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.667465 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.770514 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.770545 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.770552 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.770564 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.770575 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.873498 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.873558 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.873576 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.873599 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.873618 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.977483 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.977532 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.977549 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.977574 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:17 crc kubenswrapper[4931]: I1125 21:52:17.977591 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:17Z","lastTransitionTime":"2025-11-25T21:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.080964 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.081033 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.081066 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.081093 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.081117 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:18Z","lastTransitionTime":"2025-11-25T21:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.183977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.184037 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.184064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.184089 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.184108 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:18Z","lastTransitionTime":"2025-11-25T21:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.286644 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.286710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.286732 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.286774 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.286788 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:18Z","lastTransitionTime":"2025-11-25T21:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.390323 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.390403 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.390424 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.390449 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.390468 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:18Z","lastTransitionTime":"2025-11-25T21:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.422016 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.422056 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:18 crc kubenswrapper[4931]: E1125 21:52:18.422177 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:18 crc kubenswrapper[4931]: E1125 21:52:18.422330 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.492895 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.492946 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.492957 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.492972 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.492985 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:18Z","lastTransitionTime":"2025-11-25T21:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.595384 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.595446 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.595464 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.595488 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.595505 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:18Z","lastTransitionTime":"2025-11-25T21:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.698062 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.698119 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.698148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.698191 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.698217 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:18Z","lastTransitionTime":"2025-11-25T21:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.801423 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.801480 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.801500 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.801524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.801540 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:18Z","lastTransitionTime":"2025-11-25T21:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.903930 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.903985 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.903996 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.904016 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:18 crc kubenswrapper[4931]: I1125 21:52:18.904028 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:18Z","lastTransitionTime":"2025-11-25T21:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.006141 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.006183 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.006196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.006213 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.006226 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.108705 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.108753 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.108766 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.108783 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.108800 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.211494 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.211546 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.211563 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.211588 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.211605 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.315203 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.315256 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.315284 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.315312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.315335 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.418849 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.418952 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.418977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.419006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.419027 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.422188 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.422205 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:19 crc kubenswrapper[4931]: E1125 21:52:19.422379 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:19 crc kubenswrapper[4931]: E1125 21:52:19.422613 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.521773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.521842 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.521867 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.521928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.521951 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.625159 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.625243 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.625268 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.625301 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.625327 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.728220 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.728273 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.728288 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.728307 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.728322 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.773937 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.774014 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.774031 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.774059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.774076 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: E1125 21:52:19.794720 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:19Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.800277 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.800330 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.800342 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.800358 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.800371 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: E1125 21:52:19.819396 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:19Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.823660 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.823687 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.823699 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.823713 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.823724 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: E1125 21:52:19.842729 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:19Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.848775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.848915 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.848943 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.848972 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.848992 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: E1125 21:52:19.868602 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:19Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.872987 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.873045 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.873062 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.873091 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.873108 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: E1125 21:52:19.893770 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"80babd60-bb40-45e4-b0a0-67856b72145a\\\",\\\"systemUUID\\\":\\\"dfb0006d-ded5-40c5-a63e-512b52836fc8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:19Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:19 crc kubenswrapper[4931]: E1125 21:52:19.894127 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.897020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.897084 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.897101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.897163 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.897181 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.999788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.999903 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:19 crc kubenswrapper[4931]: I1125 21:52:19.999921 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:19.999945 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:19.999962 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:19Z","lastTransitionTime":"2025-11-25T21:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.102508 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.102602 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.102627 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.102664 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.102689 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:20Z","lastTransitionTime":"2025-11-25T21:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.206317 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.206369 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.206378 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.206399 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.206411 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:20Z","lastTransitionTime":"2025-11-25T21:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.310585 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.310651 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.310666 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.310690 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.310706 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:20Z","lastTransitionTime":"2025-11-25T21:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.414449 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.414497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.414510 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.414529 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.414541 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:20Z","lastTransitionTime":"2025-11-25T21:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.422207 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.422302 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:20 crc kubenswrapper[4931]: E1125 21:52:20.422451 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:20 crc kubenswrapper[4931]: E1125 21:52:20.422588 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.517720 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.517765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.517776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.517797 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.517808 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:20Z","lastTransitionTime":"2025-11-25T21:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.620531 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.620589 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.620605 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.620629 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.620648 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:20Z","lastTransitionTime":"2025-11-25T21:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.722657 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.722717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.722736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.722761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.722778 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:20Z","lastTransitionTime":"2025-11-25T21:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.826021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.826123 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.826141 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.826195 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.826213 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:20Z","lastTransitionTime":"2025-11-25T21:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.929462 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.929520 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.929533 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.929551 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:20 crc kubenswrapper[4931]: I1125 21:52:20.929567 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:20Z","lastTransitionTime":"2025-11-25T21:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.032048 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.032148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.032174 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.032201 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.032223 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.167793 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.167925 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.167946 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.167974 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.168004 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.271228 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.271291 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.271309 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.271332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.271351 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.374900 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.374967 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.374989 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.375021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.375042 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.422554 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.422556 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:21 crc kubenswrapper[4931]: E1125 21:52:21.422764 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:21 crc kubenswrapper[4931]: E1125 21:52:21.422929 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.477989 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.478067 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.478088 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.478112 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.478130 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.581694 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.581753 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.581772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.581798 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.581818 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.685497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.685549 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.685569 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.685592 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.685611 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.788943 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.789294 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.789432 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.789646 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.789794 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.892735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.892797 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.892815 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.892840 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.892857 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.996616 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.997103 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.997292 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.997452 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:21 crc kubenswrapper[4931]: I1125 21:52:21.997615 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:21Z","lastTransitionTime":"2025-11-25T21:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.100495 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.101365 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.101550 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.101756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.101988 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:22Z","lastTransitionTime":"2025-11-25T21:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.205345 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.205698 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.205829 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.206183 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.206315 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:22Z","lastTransitionTime":"2025-11-25T21:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.309345 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.309412 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.309435 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.309464 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.309485 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:22Z","lastTransitionTime":"2025-11-25T21:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.412133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.412185 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.412201 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.412223 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.412241 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:22Z","lastTransitionTime":"2025-11-25T21:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.421955 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:22 crc kubenswrapper[4931]: E1125 21:52:22.422087 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.422285 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:22 crc kubenswrapper[4931]: E1125 21:52:22.422413 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.441272 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.460425 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.475016 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.495086 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.514639 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.515061 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.515105 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.515121 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.515144 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.515161 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:22Z","lastTransitionTime":"2025-11-25T21:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.531609 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.548524 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.571719 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.595813 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:55Z\\\",\\\"message\\\":\\\"ccess event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1125 21:51:55.404069 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-7v4p9\\\\nI1125 21:51:55.404072 6589 services_controller.go:454] Service openshift-route-controller-manager/route-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1125 21:51:55.404080 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404086 6589 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404090 6589 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 21:51:55.404095 6589 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 21:51:55.404100 6589 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404108 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jjhc2\\\\nI1125 21:51:55.404112 6589 obj_retry.go:365] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.618546 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.618596 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.618614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.618637 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.618655 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:22Z","lastTransitionTime":"2025-11-25T21:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.619737 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.639188 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:52:14Z\\\",\\\"message\\\":\\\"2025-11-25T21:51:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449\\\\n2025-11-25T21:51:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449 to /host/opt/cni/bin/\\\\n2025-11-25T21:51:29Z [verbose] multus-daemon started\\\\n2025-11-25T21:51:29Z [verbose] Readiness Indicator file check\\\\n2025-11-25T21:52:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:52:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.657759 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.670367 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.687544 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.705765 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.721474 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.721540 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.721558 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.721584 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.721602 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:22Z","lastTransitionTime":"2025-11-25T21:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.723096 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.735494 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.746581 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:22Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.824086 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.824122 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.824134 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.824150 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.824161 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:22Z","lastTransitionTime":"2025-11-25T21:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.926584 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.926636 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.926657 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.926683 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:22 crc kubenswrapper[4931]: I1125 21:52:22.926703 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:22Z","lastTransitionTime":"2025-11-25T21:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.029569 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.029647 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.029669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.029695 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.029713 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.132559 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.132633 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.132659 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.132695 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.132722 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.235972 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.236040 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.236058 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.236086 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.236104 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.339341 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.339415 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.339434 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.339460 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.339479 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.422384 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.422442 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:23 crc kubenswrapper[4931]: E1125 21:52:23.422793 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:23 crc kubenswrapper[4931]: E1125 21:52:23.423047 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.440170 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.443987 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.444060 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.444079 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.444108 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.444126 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.547500 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.547549 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.547565 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.547587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.547605 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.650739 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.650816 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.650836 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.650861 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.650915 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.753356 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.753407 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.753423 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.753445 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.753459 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.857161 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.857230 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.857247 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.857272 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.857290 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.960944 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.961000 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.961021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.961045 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:23 crc kubenswrapper[4931]: I1125 21:52:23.961063 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:23Z","lastTransitionTime":"2025-11-25T21:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.064333 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.064403 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.064421 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.064444 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.064461 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.167865 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.168008 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.168031 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.168057 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.168073 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.271414 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.271774 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.271952 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.272458 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.272616 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.376645 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.376722 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.376746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.376779 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.376802 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.422481 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.422509 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:24 crc kubenswrapper[4931]: E1125 21:52:24.422743 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:24 crc kubenswrapper[4931]: E1125 21:52:24.423058 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.480525 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.480587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.480610 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.480641 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.480667 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.583679 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.583777 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.583800 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.583823 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.583843 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.687402 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.687467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.687482 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.687511 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.687527 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.790656 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.790743 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.790768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.790801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.790825 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.894748 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.894819 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.894837 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.894863 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.894925 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.998064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.998127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.998144 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.998170 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:24 crc kubenswrapper[4931]: I1125 21:52:24.998188 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:24Z","lastTransitionTime":"2025-11-25T21:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.101199 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.101257 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.101274 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.101300 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.101316 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:25Z","lastTransitionTime":"2025-11-25T21:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.203704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.203761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.203773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.203795 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.203812 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:25Z","lastTransitionTime":"2025-11-25T21:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.306731 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.306794 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.306811 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.306840 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.306857 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:25Z","lastTransitionTime":"2025-11-25T21:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.410271 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.410344 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.410366 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.410395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.410416 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:25Z","lastTransitionTime":"2025-11-25T21:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.422180 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.422194 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:25 crc kubenswrapper[4931]: E1125 21:52:25.422434 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:25 crc kubenswrapper[4931]: E1125 21:52:25.422601 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.513324 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.513422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.513445 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.513470 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.513488 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:25Z","lastTransitionTime":"2025-11-25T21:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.615862 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.615951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.615971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.615999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.616020 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:25Z","lastTransitionTime":"2025-11-25T21:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.718722 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.718788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.718810 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.718839 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.718862 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:25Z","lastTransitionTime":"2025-11-25T21:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.821645 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.821690 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.821702 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.821718 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.821729 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:25Z","lastTransitionTime":"2025-11-25T21:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.924829 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.924917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.924935 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.924955 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:25 crc kubenswrapper[4931]: I1125 21:52:25.924971 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:25Z","lastTransitionTime":"2025-11-25T21:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.027747 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.027816 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.027839 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.027868 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.027936 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.130824 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.130959 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.130978 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.131004 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.131023 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.255181 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.255300 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.255320 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.255342 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.255359 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.358471 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.358542 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.358568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.358600 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.358622 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.360026 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.360129 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.360207 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:30.360175982 +0000 UTC m=+148.716688141 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.360251 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.360310 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:30.360292906 +0000 UTC m=+148.716805065 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.360338 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.360430 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.360473 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:30.36045943 +0000 UTC m=+148.716971579 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.422784 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.423078 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.423303 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.423526 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.424713 4931 scope.go:117] "RemoveContainer" containerID="7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.461785 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.461835 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.461852 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.461907 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.461926 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.461951 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.462033 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.462941 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.463609 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.463657 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.463725 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:30.463703293 +0000 UTC m=+148.820215452 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.463041 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.463843 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.463862 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:52:26 crc kubenswrapper[4931]: E1125 21:52:26.463943 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:30.463925259 +0000 UTC m=+148.820437418 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.566434 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.566951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.567217 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.567430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.567704 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.672239 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.672773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.672799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.672828 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.672849 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.775233 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.775304 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.775322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.775348 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.775366 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.879086 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.879146 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.879166 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.879195 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.879217 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.982687 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.982736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.982750 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.982777 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:26 crc kubenswrapper[4931]: I1125 21:52:26.982792 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:26Z","lastTransitionTime":"2025-11-25T21:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.023931 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/2.log" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.027680 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.028149 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.056501 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04593806-2035-48fe-b66d-59d29fa47eb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5190fb6f1f9bcffaf41a2be0ec4b9ef366220439b3a5d80061a910f393372c38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a03f886d8278be42544d28620c0c8dd8ede5f2e578207a74377dc8ca11fec0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a03f886d8278be42544d28620c0c8dd8ede5f2e578207a74377dc8ca11fec0a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.076717 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.085184 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.085240 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.085253 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.085270 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.085286 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:27Z","lastTransitionTime":"2025-11-25T21:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.088822 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.103310 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.112394 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.126986 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.146534 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:55Z\\\",\\\"message\\\":\\\"ccess event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1125 21:51:55.404069 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-7v4p9\\\\nI1125 21:51:55.404072 6589 services_controller.go:454] Service openshift-route-controller-manager/route-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1125 21:51:55.404080 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404086 6589 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404090 6589 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 21:51:55.404095 6589 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 21:51:55.404100 6589 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404108 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jjhc2\\\\nI1125 21:51:55.404112 6589 obj_retry.go:365] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:52:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.165362 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.178617 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:52:14Z\\\",\\\"message\\\":\\\"2025-11-25T21:51:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449\\\\n2025-11-25T21:51:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449 to /host/opt/cni/bin/\\\\n2025-11-25T21:51:29Z [verbose] multus-daemon started\\\\n2025-11-25T21:51:29Z [verbose] Readiness Indicator file check\\\\n2025-11-25T21:52:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:52:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.187467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.187566 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.187592 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.187625 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.187650 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:27Z","lastTransitionTime":"2025-11-25T21:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.191035 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.204102 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.217665 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.229333 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.240784 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.253356 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.267292 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.282677 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.290111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.290157 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.290170 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.290187 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.290198 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:27Z","lastTransitionTime":"2025-11-25T21:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.299558 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.311747 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:27Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.392803 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.392856 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.392870 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.392906 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.392917 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:27Z","lastTransitionTime":"2025-11-25T21:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.422533 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:27 crc kubenswrapper[4931]: E1125 21:52:27.422671 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.422537 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:27 crc kubenswrapper[4931]: E1125 21:52:27.422838 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.495634 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.495715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.495744 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.495778 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.495802 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:27Z","lastTransitionTime":"2025-11-25T21:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.598645 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.598706 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.598724 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.598748 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.598768 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:27Z","lastTransitionTime":"2025-11-25T21:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.701812 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.701905 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.701926 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.701951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.701969 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:27Z","lastTransitionTime":"2025-11-25T21:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.804455 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.804521 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.804541 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.804568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.804588 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:27Z","lastTransitionTime":"2025-11-25T21:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.907126 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.907205 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.907231 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.907263 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:27 crc kubenswrapper[4931]: I1125 21:52:27.907284 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:27Z","lastTransitionTime":"2025-11-25T21:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.010491 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.010568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.010594 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.010623 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.010643 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.033470 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/3.log" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.034773 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/2.log" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.039028 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" exitCode=1 Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.039093 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.039155 4931 scope.go:117] "RemoveContainer" containerID="7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.040278 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 21:52:28 crc kubenswrapper[4931]: E1125 21:52:28.040564 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.063206 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34991482ce02346d35369eb2d7a9fc89a0f17db5370088282bb3a04c8fc8e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.084820 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.104024 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xlsqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cd30323-d7c0-47e9-b280-2d71ef753965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6388e51ebc09cc730bef2b04d6f747cfe9b23da78f47d3b6a694d0592ee6e487\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckbgf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xlsqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.113550 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.113589 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.113618 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.113636 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.113651 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.127358 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c90ad5-9f43-44c7-b122-0303c788f562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9afe63f1823910595094680f1b785237c78e2a265e16db50cbf4b7e40a1d35a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5fc7d107955b01af3a769fe9bd0e2d86cd3119b07d831c66980990a5d61e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ed5346f4651d0ae7d130626876ce8f05fa9d3aacbd73c06c3632d1f41d3964a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6235f819126338881bdd8ce12057f9d10d7ebc6da625d2a48d7be117aa10e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b61adcafdd88ef362544732baf07be11b21f578920b386c22aeef91773436338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecce6f24e67ab8e9ca6602972f661345aeed4629d84416222ce302f24dccc2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edf46ded5acb602c06c750651db0d98f60441755a74a5d26fa9905f02dd7db75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcl58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fcwz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.153323 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6be6a2b-7522-4af3-aa00-68fa970fd81c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deeb63e17d1ca4e3ea1e82b8d3a2016939ec512f0d261472a7dfddf41fbe7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:51:55Z\\\",\\\"message\\\":\\\"ccess event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1125 21:51:55.404069 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-7v4p9\\\\nI1125 21:51:55.404072 6589 services_controller.go:454] Service openshift-route-controller-manager/route-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1125 21:51:55.404080 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404086 6589 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404090 6589 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 21:51:55.404095 6589 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 21:51:55.404100 6589 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 21:51:55.404108 6589 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jjhc2\\\\nI1125 21:51:55.404112 6589 obj_retry.go:365] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:52:27Z\\\",\\\"message\\\":\\\"hub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:52:27.496324 7003 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:52:27.496864 7003 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 21:52:27.497189 7003 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 21:52:27.497230 7003 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 21:52:27.497275 7003 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 21:52:27.497290 7003 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 21:52:27.497323 7003 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 21:52:27.497354 7003 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 21:52:27.497379 7003 factory.go:656] Stopping watch factory\\\\nI1125 21:52:27.497377 7003 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 21:52:27.497399 7003 ovnkube.go:599] Stopped ovnkube\\\\nI1125 21:52:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:52:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nntdj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jjhc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.170316 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04593806-2035-48fe-b66d-59d29fa47eb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5190fb6f1f9bcffaf41a2be0ec4b9ef366220439b3a5d80061a910f393372c38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a03f886d8278be42544d28620c0c8dd8ede5f2e578207a74377dc8ca11fec0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a03f886d8278be42544d28620c0c8dd8ede5f2e578207a74377dc8ca11fec0a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.184341 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1007665b-9c9b-4206-82a5-3d26b09d809d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee1749e9183903409850e63f5ea32645531310ecca9c9a232aae09c07cd9f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d4e481688e5f5c3178a3f60735b8dc5cf2f8a16e5e27348b273c3246ca1d01e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d873c3bb6e2e10d46ce9826a7c44de9be69e5f1857b3c8966a1d9add0dec38fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.199474 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"deeb760e-2656-4416-959f-11ba1746d6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77d98ecd68299068ec4a426a828ccc029ce82c18c1d1d00e4b9b4d750683aee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ff84aa19d348f365bbe5fd272560904c5371575e86caf39efc5cbb4307a56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vdgk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4hqzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.216588 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.216640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.216656 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.216681 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.216699 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.217956 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"779e1587-ca2f-4330-91dd-7251714f63bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w2n5l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7v4p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.250020 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33b8482c-bccf-45e2-9afa-5df248636d24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fd459a81cfa3ecd1e87aebdd5148a9b816192724f3e47451f8dc58ecf98a122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c527aa6c822ce476cd38f21fce7e033f7703ef86cf1b89246dce0b253bdd976\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7837cad67b7a12059bc7ab6b640e898b88a80063be8e5afc7360f1099aa8ca0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb197fa112a69d4af0ba116ee23938fbd0aa535c3e23eddd9d58aa89e2dbae62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe78cc6cf0f711ef45799236f3c2a817f6eec681236cc44faea6cee7279e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://447e8f9287d8ae926b314568da494b28b881050616a7ee6c51308cc5bf84d5f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09094dfea14dbc449d787c45b15e53dc76ab46bd2e2d5fa189dad064e712a96b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd190f3393e68b6ccc036a01d7c3198555155e83b49a90283e4e1cb6ce9d64ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.268809 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xltbl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:52:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T21:52:14Z\\\",\\\"message\\\":\\\"2025-11-25T21:51:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449\\\\n2025-11-25T21:51:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2a978f40-de65-4f3b-a1b1-1f820255b449 to /host/opt/cni/bin/\\\\n2025-11-25T21:51:29Z [verbose] multus-daemon started\\\\n2025-11-25T21:51:29Z [verbose] Readiness Indicator file check\\\\n2025-11-25T21:52:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:52:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9btrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xltbl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.289205 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74cfff73d6b2cb26a1248c6db6034b1cd7135b4b252cea8f3162769c05c81927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.309541 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.321461 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.321553 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.321581 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.321620 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.321644 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.329365 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22448de-5122-43be-853a-081347c13cf2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b536412c0538b04717cc31493d5f3bfba0a84b31e282d610f906482d20071f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh64z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djjhg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.353027 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dee48871-9ee2-4083-a5be-427a390aaf20\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T21:51:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 21:51:15.953402 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 21:51:15.954425 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3560954706/tls.crt::/tmp/serving-cert-3560954706/tls.key\\\\\\\"\\\\nI1125 21:51:21.932431 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 21:51:21.935818 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 21:51:21.935843 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 21:51:21.935861 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 21:51:21.935870 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 21:51:21.948330 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 21:51:21.948351 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948356 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 21:51:21.948360 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 21:51:21.948363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 21:51:21.948366 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 21:51:21.948369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 21:51:21.948364 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 21:51:21.949707 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.368794 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ccbf025a2127bda50fcd76a45c3e2877a7ddbaa7945cddba4cb2c3b080ad186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528958a915dc811d424edc4237f616fd8d08fcf3163a2426e488504d9c6e02b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.384469 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfn4g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2e774b7-8bb0-4625-a50d-f69effe7ccad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdaee74c74bb92741e7b44ed4e1b70e88e3ec02b28042f1e4c396404993787be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbk6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfn4g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.405103 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"919aa0e0-3d31-4619-a30f-5f1cda16567f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7dd11fda40aa696ddeaba15ea978c7649080a892e69dc453f72d054a6ae6eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b05f6cbe0eef6f4966ce2e70f063a12c4d6a4d8c81fa73d39a0b2a9f4ded83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c1385cca08ff017b21a088740c1eaecad5c3a8986d4ca0e7c12c57bc51e9a7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T21:51:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeab02c8865e397cef8e93b1c2d199a6bc1a371c95b8942b9b4fd2282f518bd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T21:51:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T21:51:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T21:51:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.422120 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:28 crc kubenswrapper[4931]: E1125 21:52:28.422320 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.422401 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:28 crc kubenswrapper[4931]: E1125 21:52:28.423306 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.425110 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.425186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.425231 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.425267 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.425293 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.425429 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T21:51:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T21:52:28Z is after 2025-08-24T17:21:41Z" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.528456 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.528522 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.528540 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.528565 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.528584 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.630855 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.630970 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.630990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.631017 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.631035 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.734230 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.734301 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.734319 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.734343 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.734363 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.837679 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.837749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.837773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.837804 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.837827 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.941811 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.941919 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.941947 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.941977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:28 crc kubenswrapper[4931]: I1125 21:52:28.941999 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:28Z","lastTransitionTime":"2025-11-25T21:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.052782 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.052814 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.052825 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.052840 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.052851 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.056746 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/3.log" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.061342 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 21:52:29 crc kubenswrapper[4931]: E1125 21:52:29.061539 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.133918 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=67.13386343 podStartE2EDuration="1m7.13386343s" podCreationTimestamp="2025-11-25 21:51:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.111630382 +0000 UTC m=+87.468142541" watchObservedRunningTime="2025-11-25 21:52:29.13386343 +0000 UTC m=+87.490375559" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.156370 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.156415 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.156426 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.156444 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.156457 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.206731 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podStartSLOduration=66.206714992 podStartE2EDuration="1m6.206714992s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.189207433 +0000 UTC m=+87.545719592" watchObservedRunningTime="2025-11-25 21:52:29.206714992 +0000 UTC m=+87.563227111" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.206827 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=40.206824455 podStartE2EDuration="40.206824455s" podCreationTimestamp="2025-11-25 21:51:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.206578088 +0000 UTC m=+87.563090237" watchObservedRunningTime="2025-11-25 21:52:29.206824455 +0000 UTC m=+87.563336574" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.259587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.259671 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.259696 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.259730 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.259755 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.294407 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dfn4g" podStartSLOduration=66.294378359 podStartE2EDuration="1m6.294378359s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.249394369 +0000 UTC m=+87.605906508" watchObservedRunningTime="2025-11-25 21:52:29.294378359 +0000 UTC m=+87.650890528" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.309492 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=6.309458391 podStartE2EDuration="6.309458391s" podCreationTimestamp="2025-11-25 21:52:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.308536356 +0000 UTC m=+87.665048495" watchObservedRunningTime="2025-11-25 21:52:29.309458391 +0000 UTC m=+87.665970560" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.324184 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=65.324163023 podStartE2EDuration="1m5.324163023s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.323218787 +0000 UTC m=+87.679730936" watchObservedRunningTime="2025-11-25 21:52:29.324163023 +0000 UTC m=+87.680675162" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.363345 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.363388 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.363401 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.363421 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.363435 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.394628 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-xlsqn" podStartSLOduration=66.394598389 podStartE2EDuration="1m6.394598389s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.373243875 +0000 UTC m=+87.729756034" watchObservedRunningTime="2025-11-25 21:52:29.394598389 +0000 UTC m=+87.751110538" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.395669 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-fcwz5" podStartSLOduration=65.395661608 podStartE2EDuration="1m5.395661608s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.393069827 +0000 UTC m=+87.749581956" watchObservedRunningTime="2025-11-25 21:52:29.395661608 +0000 UTC m=+87.752173757" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.421702 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.421811 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:29 crc kubenswrapper[4931]: E1125 21:52:29.421924 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:29 crc kubenswrapper[4931]: E1125 21:52:29.422022 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.437475 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=63.437446101 podStartE2EDuration="1m3.437446101s" podCreationTimestamp="2025-11-25 21:51:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.436784772 +0000 UTC m=+87.793296921" watchObservedRunningTime="2025-11-25 21:52:29.437446101 +0000 UTC m=+87.793958240" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.466561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.466608 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.466621 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.466640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.466653 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.478056 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-xltbl" podStartSLOduration=65.478034421 podStartE2EDuration="1m5.478034421s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.45752215 +0000 UTC m=+87.814034309" watchObservedRunningTime="2025-11-25 21:52:29.478034421 +0000 UTC m=+87.834546550" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.478836 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4hqzf" podStartSLOduration=65.478829182 podStartE2EDuration="1m5.478829182s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:29.477148426 +0000 UTC m=+87.833660565" watchObservedRunningTime="2025-11-25 21:52:29.478829182 +0000 UTC m=+87.835341321" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.569412 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.569450 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.569462 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.569478 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.569489 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.673491 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.673549 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.673561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.673580 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.673594 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.776995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.777069 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.777088 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.777119 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.777139 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.880587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.880668 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.880691 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.880728 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.880755 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.938429 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.938497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.938516 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.938542 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.938569 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T21:52:29Z","lastTransitionTime":"2025-11-25T21:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.994949 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b"] Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.995626 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.997794 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.997816 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 21:52:29 crc kubenswrapper[4931]: I1125 21:52:29.998497 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.001293 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.104479 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66fad133-73ad-4f64-bf9b-98e58ee543c5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.104561 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/66fad133-73ad-4f64-bf9b-98e58ee543c5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.104603 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66fad133-73ad-4f64-bf9b-98e58ee543c5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.104831 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/66fad133-73ad-4f64-bf9b-98e58ee543c5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.104964 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66fad133-73ad-4f64-bf9b-98e58ee543c5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.206196 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/66fad133-73ad-4f64-bf9b-98e58ee543c5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.206328 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66fad133-73ad-4f64-bf9b-98e58ee543c5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.206343 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/66fad133-73ad-4f64-bf9b-98e58ee543c5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.206369 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66fad133-73ad-4f64-bf9b-98e58ee543c5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.206431 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/66fad133-73ad-4f64-bf9b-98e58ee543c5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.206472 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66fad133-73ad-4f64-bf9b-98e58ee543c5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.206548 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/66fad133-73ad-4f64-bf9b-98e58ee543c5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.208364 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66fad133-73ad-4f64-bf9b-98e58ee543c5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.219234 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66fad133-73ad-4f64-bf9b-98e58ee543c5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.229975 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66fad133-73ad-4f64-bf9b-98e58ee543c5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p664b\" (UID: \"66fad133-73ad-4f64-bf9b-98e58ee543c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.318193 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" Nov 25 21:52:30 crc kubenswrapper[4931]: W1125 21:52:30.350926 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66fad133_73ad_4f64_bf9b_98e58ee543c5.slice/crio-0bb633e4d5c575ccd19be4e4cb1b8ee8a0646a2a95cc520984536f40307c4bf6 WatchSource:0}: Error finding container 0bb633e4d5c575ccd19be4e4cb1b8ee8a0646a2a95cc520984536f40307c4bf6: Status 404 returned error can't find the container with id 0bb633e4d5c575ccd19be4e4cb1b8ee8a0646a2a95cc520984536f40307c4bf6 Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.422574 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:30 crc kubenswrapper[4931]: I1125 21:52:30.422747 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:30 crc kubenswrapper[4931]: E1125 21:52:30.423010 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:30 crc kubenswrapper[4931]: E1125 21:52:30.423270 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:31 crc kubenswrapper[4931]: I1125 21:52:31.068795 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" event={"ID":"66fad133-73ad-4f64-bf9b-98e58ee543c5","Type":"ContainerStarted","Data":"06bbb55c08f98fd69bb9f6db58e657b7f71fcd0e92cc3cf2d5c0b34e8935798f"} Nov 25 21:52:31 crc kubenswrapper[4931]: I1125 21:52:31.068845 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" event={"ID":"66fad133-73ad-4f64-bf9b-98e58ee543c5","Type":"ContainerStarted","Data":"0bb633e4d5c575ccd19be4e4cb1b8ee8a0646a2a95cc520984536f40307c4bf6"} Nov 25 21:52:31 crc kubenswrapper[4931]: I1125 21:52:31.093407 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p664b" podStartSLOduration=68.093373697 podStartE2EDuration="1m8.093373697s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:52:31.090245161 +0000 UTC m=+89.446757310" watchObservedRunningTime="2025-11-25 21:52:31.093373697 +0000 UTC m=+89.449885856" Nov 25 21:52:31 crc kubenswrapper[4931]: I1125 21:52:31.421833 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:31 crc kubenswrapper[4931]: I1125 21:52:31.421833 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:31 crc kubenswrapper[4931]: E1125 21:52:31.422061 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:31 crc kubenswrapper[4931]: E1125 21:52:31.422159 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:32 crc kubenswrapper[4931]: I1125 21:52:32.422491 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:32 crc kubenswrapper[4931]: I1125 21:52:32.424707 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:32 crc kubenswrapper[4931]: E1125 21:52:32.424907 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:32 crc kubenswrapper[4931]: E1125 21:52:32.425222 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:33 crc kubenswrapper[4931]: I1125 21:52:33.421749 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:33 crc kubenswrapper[4931]: I1125 21:52:33.421785 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:33 crc kubenswrapper[4931]: E1125 21:52:33.422053 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:33 crc kubenswrapper[4931]: E1125 21:52:33.422415 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:34 crc kubenswrapper[4931]: I1125 21:52:34.422413 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:34 crc kubenswrapper[4931]: I1125 21:52:34.422436 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:34 crc kubenswrapper[4931]: E1125 21:52:34.422630 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:34 crc kubenswrapper[4931]: E1125 21:52:34.422697 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:35 crc kubenswrapper[4931]: I1125 21:52:35.421925 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:35 crc kubenswrapper[4931]: I1125 21:52:35.421942 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:35 crc kubenswrapper[4931]: E1125 21:52:35.422116 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:35 crc kubenswrapper[4931]: E1125 21:52:35.422192 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:36 crc kubenswrapper[4931]: I1125 21:52:36.422592 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:36 crc kubenswrapper[4931]: I1125 21:52:36.422667 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:36 crc kubenswrapper[4931]: E1125 21:52:36.422797 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:36 crc kubenswrapper[4931]: E1125 21:52:36.422960 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:37 crc kubenswrapper[4931]: I1125 21:52:37.421801 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:37 crc kubenswrapper[4931]: E1125 21:52:37.422040 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:37 crc kubenswrapper[4931]: I1125 21:52:37.422066 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:37 crc kubenswrapper[4931]: E1125 21:52:37.422447 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:38 crc kubenswrapper[4931]: I1125 21:52:38.422649 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:38 crc kubenswrapper[4931]: I1125 21:52:38.422745 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:38 crc kubenswrapper[4931]: E1125 21:52:38.422825 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:38 crc kubenswrapper[4931]: E1125 21:52:38.423032 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:39 crc kubenswrapper[4931]: I1125 21:52:39.422435 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:39 crc kubenswrapper[4931]: I1125 21:52:39.422501 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:39 crc kubenswrapper[4931]: E1125 21:52:39.422674 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:39 crc kubenswrapper[4931]: E1125 21:52:39.423283 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:40 crc kubenswrapper[4931]: I1125 21:52:40.422105 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:40 crc kubenswrapper[4931]: E1125 21:52:40.422295 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:40 crc kubenswrapper[4931]: I1125 21:52:40.422137 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:40 crc kubenswrapper[4931]: E1125 21:52:40.422686 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:40 crc kubenswrapper[4931]: I1125 21:52:40.423830 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 21:52:40 crc kubenswrapper[4931]: E1125 21:52:40.424177 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:52:41 crc kubenswrapper[4931]: I1125 21:52:41.422162 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:41 crc kubenswrapper[4931]: I1125 21:52:41.422206 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:41 crc kubenswrapper[4931]: E1125 21:52:41.422472 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:41 crc kubenswrapper[4931]: E1125 21:52:41.422694 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:42 crc kubenswrapper[4931]: I1125 21:52:42.346778 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:42 crc kubenswrapper[4931]: E1125 21:52:42.347023 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:52:42 crc kubenswrapper[4931]: E1125 21:52:42.347114 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs podName:779e1587-ca2f-4330-91dd-7251714f63bc nodeName:}" failed. No retries permitted until 2025-11-25 21:53:46.347090973 +0000 UTC m=+164.703603122 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs") pod "network-metrics-daemon-7v4p9" (UID: "779e1587-ca2f-4330-91dd-7251714f63bc") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 21:52:42 crc kubenswrapper[4931]: I1125 21:52:42.421575 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:42 crc kubenswrapper[4931]: I1125 21:52:42.421664 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:42 crc kubenswrapper[4931]: E1125 21:52:42.423715 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:42 crc kubenswrapper[4931]: E1125 21:52:42.423963 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:43 crc kubenswrapper[4931]: I1125 21:52:43.422369 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:43 crc kubenswrapper[4931]: E1125 21:52:43.422553 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:43 crc kubenswrapper[4931]: I1125 21:52:43.422411 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:43 crc kubenswrapper[4931]: E1125 21:52:43.422763 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:44 crc kubenswrapper[4931]: I1125 21:52:44.422301 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:44 crc kubenswrapper[4931]: I1125 21:52:44.422365 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:44 crc kubenswrapper[4931]: E1125 21:52:44.422545 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:44 crc kubenswrapper[4931]: E1125 21:52:44.422681 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:45 crc kubenswrapper[4931]: I1125 21:52:45.422615 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:45 crc kubenswrapper[4931]: I1125 21:52:45.422693 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:45 crc kubenswrapper[4931]: E1125 21:52:45.423269 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:45 crc kubenswrapper[4931]: E1125 21:52:45.423389 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:46 crc kubenswrapper[4931]: I1125 21:52:46.422480 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:46 crc kubenswrapper[4931]: I1125 21:52:46.422578 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:46 crc kubenswrapper[4931]: E1125 21:52:46.422698 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:46 crc kubenswrapper[4931]: E1125 21:52:46.422839 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:47 crc kubenswrapper[4931]: I1125 21:52:47.421966 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:47 crc kubenswrapper[4931]: I1125 21:52:47.422253 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:47 crc kubenswrapper[4931]: E1125 21:52:47.422534 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:47 crc kubenswrapper[4931]: E1125 21:52:47.422986 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:48 crc kubenswrapper[4931]: I1125 21:52:48.426790 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:48 crc kubenswrapper[4931]: E1125 21:52:48.427025 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:48 crc kubenswrapper[4931]: I1125 21:52:48.427179 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:48 crc kubenswrapper[4931]: E1125 21:52:48.427348 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:49 crc kubenswrapper[4931]: I1125 21:52:49.422303 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:49 crc kubenswrapper[4931]: I1125 21:52:49.422360 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:49 crc kubenswrapper[4931]: E1125 21:52:49.422511 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:49 crc kubenswrapper[4931]: E1125 21:52:49.422648 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:50 crc kubenswrapper[4931]: I1125 21:52:50.422507 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:50 crc kubenswrapper[4931]: I1125 21:52:50.422518 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:50 crc kubenswrapper[4931]: E1125 21:52:50.422748 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:50 crc kubenswrapper[4931]: E1125 21:52:50.422867 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:51 crc kubenswrapper[4931]: I1125 21:52:51.422405 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:51 crc kubenswrapper[4931]: I1125 21:52:51.422505 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:51 crc kubenswrapper[4931]: E1125 21:52:51.422582 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:51 crc kubenswrapper[4931]: E1125 21:52:51.422703 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:51 crc kubenswrapper[4931]: I1125 21:52:51.423787 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 21:52:51 crc kubenswrapper[4931]: E1125 21:52:51.424067 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:52:52 crc kubenswrapper[4931]: I1125 21:52:52.422727 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:52 crc kubenswrapper[4931]: I1125 21:52:52.422779 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:52 crc kubenswrapper[4931]: E1125 21:52:52.424742 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:52 crc kubenswrapper[4931]: E1125 21:52:52.424946 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:53 crc kubenswrapper[4931]: I1125 21:52:53.422551 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:53 crc kubenswrapper[4931]: I1125 21:52:53.422578 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:53 crc kubenswrapper[4931]: E1125 21:52:53.422746 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:53 crc kubenswrapper[4931]: E1125 21:52:53.423045 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:54 crc kubenswrapper[4931]: I1125 21:52:54.422421 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:54 crc kubenswrapper[4931]: I1125 21:52:54.422475 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:54 crc kubenswrapper[4931]: E1125 21:52:54.422606 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:54 crc kubenswrapper[4931]: E1125 21:52:54.423010 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:55 crc kubenswrapper[4931]: I1125 21:52:55.421765 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:55 crc kubenswrapper[4931]: E1125 21:52:55.421973 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:55 crc kubenswrapper[4931]: I1125 21:52:55.421790 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:55 crc kubenswrapper[4931]: E1125 21:52:55.422660 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:56 crc kubenswrapper[4931]: I1125 21:52:56.421495 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:56 crc kubenswrapper[4931]: I1125 21:52:56.421533 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:56 crc kubenswrapper[4931]: E1125 21:52:56.421650 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:56 crc kubenswrapper[4931]: E1125 21:52:56.421920 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:57 crc kubenswrapper[4931]: I1125 21:52:57.421613 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:57 crc kubenswrapper[4931]: I1125 21:52:57.421613 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:57 crc kubenswrapper[4931]: E1125 21:52:57.421841 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:52:57 crc kubenswrapper[4931]: E1125 21:52:57.421990 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:58 crc kubenswrapper[4931]: I1125 21:52:58.422192 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:52:58 crc kubenswrapper[4931]: I1125 21:52:58.422217 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:52:58 crc kubenswrapper[4931]: E1125 21:52:58.423393 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:52:58 crc kubenswrapper[4931]: E1125 21:52:58.423421 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:52:59 crc kubenswrapper[4931]: I1125 21:52:59.422460 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:52:59 crc kubenswrapper[4931]: I1125 21:52:59.422516 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:52:59 crc kubenswrapper[4931]: E1125 21:52:59.422634 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:52:59 crc kubenswrapper[4931]: E1125 21:52:59.423015 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:00 crc kubenswrapper[4931]: I1125 21:53:00.421764 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:00 crc kubenswrapper[4931]: I1125 21:53:00.421767 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:00 crc kubenswrapper[4931]: E1125 21:53:00.421988 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:00 crc kubenswrapper[4931]: E1125 21:53:00.422227 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:01 crc kubenswrapper[4931]: I1125 21:53:01.187867 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/1.log" Nov 25 21:53:01 crc kubenswrapper[4931]: I1125 21:53:01.188717 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/0.log" Nov 25 21:53:01 crc kubenswrapper[4931]: I1125 21:53:01.188812 4931 generic.go:334] "Generic (PLEG): container finished" podID="ef281bd7-e1b1-41cd-b0a2-fa026829cf8f" containerID="0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89" exitCode=1 Nov 25 21:53:01 crc kubenswrapper[4931]: I1125 21:53:01.188900 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xltbl" event={"ID":"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f","Type":"ContainerDied","Data":"0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89"} Nov 25 21:53:01 crc kubenswrapper[4931]: I1125 21:53:01.188983 4931 scope.go:117] "RemoveContainer" containerID="fd4cb3690cd638ea678206bf4ad5c2f1871856f9a6f3c2947bb0a05eeade22e4" Nov 25 21:53:01 crc kubenswrapper[4931]: I1125 21:53:01.189502 4931 scope.go:117] "RemoveContainer" containerID="0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89" Nov 25 21:53:01 crc kubenswrapper[4931]: E1125 21:53:01.189766 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-xltbl_openshift-multus(ef281bd7-e1b1-41cd-b0a2-fa026829cf8f)\"" pod="openshift-multus/multus-xltbl" podUID="ef281bd7-e1b1-41cd-b0a2-fa026829cf8f" Nov 25 21:53:01 crc kubenswrapper[4931]: I1125 21:53:01.422238 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:01 crc kubenswrapper[4931]: I1125 21:53:01.422268 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:01 crc kubenswrapper[4931]: E1125 21:53:01.422401 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:01 crc kubenswrapper[4931]: E1125 21:53:01.422514 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:02 crc kubenswrapper[4931]: I1125 21:53:02.195110 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/1.log" Nov 25 21:53:02 crc kubenswrapper[4931]: E1125 21:53:02.370634 4931 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 25 21:53:02 crc kubenswrapper[4931]: I1125 21:53:02.422542 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:02 crc kubenswrapper[4931]: I1125 21:53:02.422651 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:02 crc kubenswrapper[4931]: E1125 21:53:02.424368 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:02 crc kubenswrapper[4931]: E1125 21:53:02.424563 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:02 crc kubenswrapper[4931]: E1125 21:53:02.549028 4931 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 21:53:03 crc kubenswrapper[4931]: I1125 21:53:03.421937 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:03 crc kubenswrapper[4931]: I1125 21:53:03.421949 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:03 crc kubenswrapper[4931]: E1125 21:53:03.422157 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:03 crc kubenswrapper[4931]: E1125 21:53:03.422363 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:04 crc kubenswrapper[4931]: I1125 21:53:04.422497 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:04 crc kubenswrapper[4931]: E1125 21:53:04.422692 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:04 crc kubenswrapper[4931]: I1125 21:53:04.423101 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:04 crc kubenswrapper[4931]: E1125 21:53:04.423254 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:05 crc kubenswrapper[4931]: I1125 21:53:05.422630 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:05 crc kubenswrapper[4931]: I1125 21:53:05.422747 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:05 crc kubenswrapper[4931]: E1125 21:53:05.422843 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:05 crc kubenswrapper[4931]: E1125 21:53:05.423004 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:05 crc kubenswrapper[4931]: I1125 21:53:05.424503 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 21:53:05 crc kubenswrapper[4931]: E1125 21:53:05.425039 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jjhc2_openshift-ovn-kubernetes(b6be6a2b-7522-4af3-aa00-68fa970fd81c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" Nov 25 21:53:06 crc kubenswrapper[4931]: I1125 21:53:06.421935 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:06 crc kubenswrapper[4931]: E1125 21:53:06.422127 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:06 crc kubenswrapper[4931]: I1125 21:53:06.422233 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:06 crc kubenswrapper[4931]: E1125 21:53:06.422433 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:07 crc kubenswrapper[4931]: I1125 21:53:07.422049 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:07 crc kubenswrapper[4931]: I1125 21:53:07.422088 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:07 crc kubenswrapper[4931]: E1125 21:53:07.422727 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:07 crc kubenswrapper[4931]: E1125 21:53:07.422935 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:07 crc kubenswrapper[4931]: E1125 21:53:07.550530 4931 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 21:53:08 crc kubenswrapper[4931]: I1125 21:53:08.421991 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:08 crc kubenswrapper[4931]: I1125 21:53:08.422050 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:08 crc kubenswrapper[4931]: E1125 21:53:08.422184 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:08 crc kubenswrapper[4931]: E1125 21:53:08.422366 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:09 crc kubenswrapper[4931]: I1125 21:53:09.422252 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:09 crc kubenswrapper[4931]: I1125 21:53:09.422268 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:09 crc kubenswrapper[4931]: E1125 21:53:09.422965 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:09 crc kubenswrapper[4931]: E1125 21:53:09.423157 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:10 crc kubenswrapper[4931]: I1125 21:53:10.422183 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:10 crc kubenswrapper[4931]: E1125 21:53:10.422413 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:10 crc kubenswrapper[4931]: I1125 21:53:10.422516 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:10 crc kubenswrapper[4931]: E1125 21:53:10.422716 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:11 crc kubenswrapper[4931]: I1125 21:53:11.422489 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:11 crc kubenswrapper[4931]: I1125 21:53:11.422608 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:11 crc kubenswrapper[4931]: E1125 21:53:11.422776 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:11 crc kubenswrapper[4931]: E1125 21:53:11.423024 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:12 crc kubenswrapper[4931]: I1125 21:53:12.423193 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:12 crc kubenswrapper[4931]: I1125 21:53:12.423325 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:12 crc kubenswrapper[4931]: E1125 21:53:12.425050 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:12 crc kubenswrapper[4931]: E1125 21:53:12.425275 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:12 crc kubenswrapper[4931]: E1125 21:53:12.552267 4931 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 21:53:13 crc kubenswrapper[4931]: I1125 21:53:13.421845 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:13 crc kubenswrapper[4931]: I1125 21:53:13.422103 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:13 crc kubenswrapper[4931]: E1125 21:53:13.422396 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:13 crc kubenswrapper[4931]: E1125 21:53:13.422680 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:14 crc kubenswrapper[4931]: I1125 21:53:14.421981 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:14 crc kubenswrapper[4931]: I1125 21:53:14.422166 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:14 crc kubenswrapper[4931]: I1125 21:53:14.422317 4931 scope.go:117] "RemoveContainer" containerID="0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89" Nov 25 21:53:14 crc kubenswrapper[4931]: E1125 21:53:14.422475 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:14 crc kubenswrapper[4931]: E1125 21:53:14.422630 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:15 crc kubenswrapper[4931]: I1125 21:53:15.247402 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/1.log" Nov 25 21:53:15 crc kubenswrapper[4931]: I1125 21:53:15.247504 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xltbl" event={"ID":"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f","Type":"ContainerStarted","Data":"18799a6d8db1c654581e5b8b65ff9892c6445ce545d95d10a08db229ac8c7296"} Nov 25 21:53:15 crc kubenswrapper[4931]: I1125 21:53:15.422032 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:15 crc kubenswrapper[4931]: I1125 21:53:15.422104 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:15 crc kubenswrapper[4931]: E1125 21:53:15.422173 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:15 crc kubenswrapper[4931]: E1125 21:53:15.422285 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:16 crc kubenswrapper[4931]: I1125 21:53:16.422631 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:16 crc kubenswrapper[4931]: E1125 21:53:16.422954 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:16 crc kubenswrapper[4931]: I1125 21:53:16.423050 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:16 crc kubenswrapper[4931]: E1125 21:53:16.423363 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:16 crc kubenswrapper[4931]: I1125 21:53:16.423729 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 21:53:17 crc kubenswrapper[4931]: I1125 21:53:17.256568 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/3.log" Nov 25 21:53:17 crc kubenswrapper[4931]: I1125 21:53:17.259574 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerStarted","Data":"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9"} Nov 25 21:53:17 crc kubenswrapper[4931]: I1125 21:53:17.422209 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:17 crc kubenswrapper[4931]: I1125 21:53:17.422311 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:17 crc kubenswrapper[4931]: E1125 21:53:17.422414 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:17 crc kubenswrapper[4931]: E1125 21:53:17.422650 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:17 crc kubenswrapper[4931]: E1125 21:53:17.553432 4931 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 21:53:17 crc kubenswrapper[4931]: I1125 21:53:17.951338 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7v4p9"] Nov 25 21:53:18 crc kubenswrapper[4931]: I1125 21:53:18.262258 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:18 crc kubenswrapper[4931]: E1125 21:53:18.262385 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:18 crc kubenswrapper[4931]: I1125 21:53:18.422410 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:18 crc kubenswrapper[4931]: I1125 21:53:18.422444 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:18 crc kubenswrapper[4931]: E1125 21:53:18.422627 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:18 crc kubenswrapper[4931]: E1125 21:53:18.422667 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:19 crc kubenswrapper[4931]: I1125 21:53:19.422647 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:19 crc kubenswrapper[4931]: E1125 21:53:19.422937 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:20 crc kubenswrapper[4931]: I1125 21:53:20.422446 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:20 crc kubenswrapper[4931]: I1125 21:53:20.422559 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:20 crc kubenswrapper[4931]: E1125 21:53:20.422685 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:20 crc kubenswrapper[4931]: I1125 21:53:20.422561 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:20 crc kubenswrapper[4931]: E1125 21:53:20.422868 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:20 crc kubenswrapper[4931]: E1125 21:53:20.423097 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:21 crc kubenswrapper[4931]: I1125 21:53:21.422616 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:21 crc kubenswrapper[4931]: E1125 21:53:21.422850 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 21:53:22 crc kubenswrapper[4931]: I1125 21:53:22.421906 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:22 crc kubenswrapper[4931]: I1125 21:53:22.422030 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:22 crc kubenswrapper[4931]: I1125 21:53:22.422098 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:22 crc kubenswrapper[4931]: E1125 21:53:22.424015 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 21:53:22 crc kubenswrapper[4931]: E1125 21:53:22.424803 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7v4p9" podUID="779e1587-ca2f-4330-91dd-7251714f63bc" Nov 25 21:53:22 crc kubenswrapper[4931]: E1125 21:53:22.424930 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 21:53:23 crc kubenswrapper[4931]: I1125 21:53:23.422501 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:23 crc kubenswrapper[4931]: I1125 21:53:23.425385 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 21:53:23 crc kubenswrapper[4931]: I1125 21:53:23.425596 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 21:53:24 crc kubenswrapper[4931]: I1125 21:53:24.422001 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:24 crc kubenswrapper[4931]: I1125 21:53:24.422028 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:24 crc kubenswrapper[4931]: I1125 21:53:24.422166 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:24 crc kubenswrapper[4931]: I1125 21:53:24.424701 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 21:53:24 crc kubenswrapper[4931]: I1125 21:53:24.425068 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 21:53:24 crc kubenswrapper[4931]: I1125 21:53:24.425297 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 21:53:24 crc kubenswrapper[4931]: I1125 21:53:24.426216 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.398549 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:30 crc kubenswrapper[4931]: E1125 21:53:30.398829 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:55:32.398789634 +0000 UTC m=+270.755301793 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.399117 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.399180 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.400615 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.415628 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.446390 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.500749 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.501138 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.504371 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.504991 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.646152 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.661352 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podStartSLOduration=126.661334465 podStartE2EDuration="2m6.661334465s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:18.286047066 +0000 UTC m=+136.642559195" watchObservedRunningTime="2025-11-25 21:53:30.661334465 +0000 UTC m=+149.017846584" Nov 25 21:53:30 crc kubenswrapper[4931]: W1125 21:53:30.670416 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-1136ccc97892151b214c61df562094a037ca9768b55ed63f63b1fa869343f7dc WatchSource:0}: Error finding container 1136ccc97892151b214c61df562094a037ca9768b55ed63f63b1fa869343f7dc: Status 404 returned error can't find the container with id 1136ccc97892151b214c61df562094a037ca9768b55ed63f63b1fa869343f7dc Nov 25 21:53:30 crc kubenswrapper[4931]: I1125 21:53:30.772825 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:30 crc kubenswrapper[4931]: W1125 21:53:30.821647 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-545875266e2308f3c8d77e92a81d194f190082294d554a2dc3fe9d8f6376318a WatchSource:0}: Error finding container 545875266e2308f3c8d77e92a81d194f190082294d554a2dc3fe9d8f6376318a: Status 404 returned error can't find the container with id 545875266e2308f3c8d77e92a81d194f190082294d554a2dc3fe9d8f6376318a Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.001222 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-fa226fd14ae14ef989a61b23ec1dae0e4b9dee80a19a6ff932c9be72a4c6b8d5 WatchSource:0}: Error finding container fa226fd14ae14ef989a61b23ec1dae0e4b9dee80a19a6ff932c9be72a4c6b8d5: Status 404 returned error can't find the container with id fa226fd14ae14ef989a61b23ec1dae0e4b9dee80a19a6ff932c9be72a4c6b8d5 Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.321350 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9a745a53f83abedda749377fef46c88bdadb75440a261a6a6a1364a24b1068cd"} Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.321416 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1136ccc97892151b214c61df562094a037ca9768b55ed63f63b1fa869343f7dc"} Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.327104 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"4e8b227e5a480cc0c7529d65452779f24c6a3fe8d253906f0e396e58356cc139"} Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.327180 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"545875266e2308f3c8d77e92a81d194f190082294d554a2dc3fe9d8f6376318a"} Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.329133 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ecc236ea660a4ea6a8a0a911fcfbbe5e0d1ab8683442f9da4e81a3b3bbe88174"} Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.329194 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"fa226fd14ae14ef989a61b23ec1dae0e4b9dee80a19a6ff932c9be72a4c6b8d5"} Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.329374 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.363372 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.455657 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.456146 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.457719 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.458492 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.459323 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.459596 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.459678 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.459855 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.460331 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.460533 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.461393 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kvz7t"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.461769 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.462000 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.462579 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.463197 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.463560 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.463915 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.464272 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vqmcg"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.464511 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.464866 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.464956 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.465054 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.468630 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p4p2g"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.469211 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.472975 4931 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-serving-cert": failed to list *v1.Secret: secrets "v4-0-config-system-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.473017 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.473062 4931 reflector.go:561] object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc": failed to list *v1.Secret: secrets "oauth-openshift-dockercfg-znhcc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.473077 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-znhcc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-openshift-dockercfg-znhcc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.473113 4931 reflector.go:561] object-"openshift-console-operator"/"trusted-ca": failed to list *v1.ConfigMap: configmaps "trusted-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.473123 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-console-operator\"/\"trusted-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.473162 4931 reflector.go:561] object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr": failed to list *v1.Secret: secrets "console-operator-dockercfg-4xjcr" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.473172 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-console-operator\"/\"console-operator-dockercfg-4xjcr\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"console-operator-dockercfg-4xjcr\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.473238 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.473615 4931 reflector.go:561] object-"openshift-console-operator"/"console-operator-config": failed to list *v1.ConfigMap: configmaps "console-operator-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.473713 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-console-operator\"/\"console-operator-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"console-operator-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.474108 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.474290 4931 reflector.go:561] object-"openshift-controller-manager-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.474405 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.474417 4931 reflector.go:561] object-"openshift-console-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.474532 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.474448 4931 reflector.go:561] object-"openshift-controller-manager-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.474640 4931 reflector.go:561] object-"openshift-route-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.474669 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.474715 4931 reflector.go:561] object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert": failed to list *v1.Secret: secrets "openshift-controller-manager-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.474730 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.474791 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.474648 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.474936 4931 reflector.go:561] object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2": failed to list *v1.Secret: secrets "route-controller-manager-sa-dockercfg-h2zr2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.474959 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-h2zr2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"route-controller-manager-sa-dockercfg-h2zr2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.474474 4931 reflector.go:561] object-"openshift-route-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.474986 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475049 4931 reflector.go:561] object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw": failed to list *v1.Secret: secrets "openshift-controller-manager-operator-dockercfg-vw8fw" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475064 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-dockercfg-vw8fw\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-operator-dockercfg-vw8fw\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475114 4931 reflector.go:561] object-"openshift-route-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475128 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475168 4931 reflector.go:561] object-"openshift-console-operator"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475182 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-console-operator\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475223 4931 reflector.go:561] object-"openshift-route-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475236 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475277 4931 reflector.go:561] object-"openshift-console-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475289 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-console-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475328 4931 reflector.go:561] object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config": failed to list *v1.ConfigMap: configmaps "openshift-controller-manager-operator-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager-operator": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475320 4931 reflector.go:561] object-"openshift-route-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475374 4931 reflector.go:561] object-"openshift-authentication"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475388 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475387 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475427 4931 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data": failed to list *v1.Secret: secrets "v4-0-config-user-idp-0-file-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475444 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-idp-0-file-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475481 4931 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-session": failed to list *v1.Secret: secrets "v4-0-config-system-session" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475493 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-session\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-session\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475342 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-controller-manager-operator-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.475559 4931 reflector.go:561] object-"openshift-authentication"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.475589 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.475564 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.477421 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.478285 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.478639 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: W1125 21:53:31.483111 4931 reflector.go:561] object-"openshift-authentication"/"audit": failed to list *v1.ConfigMap: configmaps "audit" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Nov 25 21:53:31 crc kubenswrapper[4931]: E1125 21:53:31.483149 4931 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"audit\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.486938 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.487623 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.487777 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-2hpxq"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.488330 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.489108 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.489399 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.489595 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.489616 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.489621 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.490061 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.491536 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.491919 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.492096 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.492202 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.493781 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.494017 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.494193 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.494369 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.497961 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.498313 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.498390 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.498653 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.498771 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.498950 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.499103 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.499140 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.499002 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.499667 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.499807 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.500104 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.500212 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.500313 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.500420 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.500952 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.501094 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.508997 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f8tsr"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.509596 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-wvj9s"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.509923 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.510069 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.512720 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-9dxf2"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.521327 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9dxf2" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.525824 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.546041 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5dlk4"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.546787 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.546956 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.547817 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.548396 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.553395 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.553605 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.554221 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.554364 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.554494 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.554629 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.554709 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.555104 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.555681 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.555757 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.555779 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.555826 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.555966 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.556045 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.556132 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.556148 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.556149 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.556242 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.556286 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.556295 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.556340 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.557137 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.558976 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.559420 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.561011 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.561168 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-84p7l"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.561508 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.561686 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.562109 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.562521 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.562717 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.563208 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.563360 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.563463 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.563716 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.565229 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.568215 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hsz8c"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.568866 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.569084 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.569139 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.569147 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.570487 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.571002 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.571832 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.572916 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.573258 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.573664 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.574424 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.586004 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.587637 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.593808 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.595008 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6vpzc"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.595825 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.601000 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.602738 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.614339 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.614972 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.615758 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.616518 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.617538 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.619045 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fzvcr"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.619229 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.619436 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kvz7t"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.619529 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.620188 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s2td2"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.621352 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.623838 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-image-import-ca\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.623870 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.623914 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3151332c-78d1-439b-bf22-25e60ed60dd9-serving-cert\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.623937 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3151332c-78d1-439b-bf22-25e60ed60dd9-audit-dir\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.623957 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624030 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl8hb\" (UniqueName: \"kubernetes.io/projected/ffaf251e-ab8f-4339-905d-5bd8f700beee-kube-api-access-fl8hb\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624072 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55263751-50ce-4c6b-a3e9-25c7fee34e37-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kxgml\" (UID: \"55263751-50ce-4c6b-a3e9-25c7fee34e37\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624094 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-etcd-serving-ca\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624111 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624146 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-config\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624183 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-dir\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624217 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrfln\" (UniqueName: \"kubernetes.io/projected/3151332c-78d1-439b-bf22-25e60ed60dd9-kube-api-access-zrfln\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624239 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q6np\" (UniqueName: \"kubernetes.io/projected/b4dd52d7-4597-4159-a728-012cd2180b35-kube-api-access-9q6np\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624271 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9d1e5baf-648e-4327-afc5-7a2835d7fb21-encryption-config\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624295 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-client-ca\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624314 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-console-serving-cert\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624338 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/07b1ae10-2353-4d92-850b-2370581c674f-service-ca-bundle\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624357 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srckn\" (UniqueName: \"kubernetes.io/projected/74cbb6cf-4717-439c-b712-67718c154742-kube-api-access-srckn\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624374 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fada681c-6ade-49e4-b324-1985853bc61e-serving-cert\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624391 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/681adfb6-f6a8-4934-ade0-a1499bf95d2f-auth-proxy-config\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624414 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624432 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-config\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624463 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9d1e5baf-648e-4327-afc5-7a2835d7fb21-audit-dir\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624514 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-client-ca\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624547 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d1e5baf-648e-4327-afc5-7a2835d7fb21-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624567 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/07b1ae10-2353-4d92-850b-2370581c674f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624588 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-console-config\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624609 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-trusted-ca-bundle\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624633 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624655 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9d1e5baf-648e-4327-afc5-7a2835d7fb21-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624679 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-config\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624718 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624741 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624769 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624790 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07b1ae10-2353-4d92-850b-2370581c674f-serving-cert\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624811 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8dqx\" (UniqueName: \"kubernetes.io/projected/54f4b8f7-8d9c-4bef-af7f-cb2335639e50-kube-api-access-d8dqx\") pod \"cluster-samples-operator-665b6dd947-f5nxk\" (UID: \"54f4b8f7-8d9c-4bef-af7f-cb2335639e50\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624833 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/54f4b8f7-8d9c-4bef-af7f-cb2335639e50-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-f5nxk\" (UID: \"54f4b8f7-8d9c-4bef-af7f-cb2335639e50\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624857 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624899 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ffaf251e-ab8f-4339-905d-5bd8f700beee-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624923 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w26t\" (UniqueName: \"kubernetes.io/projected/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-kube-api-access-5w26t\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624944 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x2px\" (UniqueName: \"kubernetes.io/projected/fada681c-6ade-49e4-b324-1985853bc61e-kube-api-access-8x2px\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624971 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.624993 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625035 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3151332c-78d1-439b-bf22-25e60ed60dd9-encryption-config\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625056 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74cbb6cf-4717-439c-b712-67718c154742-serving-cert\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625078 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7zlq\" (UniqueName: \"kubernetes.io/projected/55263751-50ce-4c6b-a3e9-25c7fee34e37-kube-api-access-v7zlq\") pod \"openshift-apiserver-operator-796bbdcf4f-kxgml\" (UID: \"55263751-50ce-4c6b-a3e9-25c7fee34e37\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625110 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jrcf\" (UniqueName: \"kubernetes.io/projected/681adfb6-f6a8-4934-ade0-a1499bf95d2f-kube-api-access-9jrcf\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625131 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625153 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d1e5baf-648e-4327-afc5-7a2835d7fb21-serving-cert\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625176 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625197 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55263751-50ce-4c6b-a3e9-25c7fee34e37-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kxgml\" (UID: \"55263751-50ce-4c6b-a3e9-25c7fee34e37\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625230 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffaf251e-ab8f-4339-905d-5bd8f700beee-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625256 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-oauth-serving-cert\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625274 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-audit\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625293 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4dd52d7-4597-4159-a728-012cd2180b35-serving-cert\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625313 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625330 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625347 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-config\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625363 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625378 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ffaf251e-ab8f-4339-905d-5bd8f700beee-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625397 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4qmx\" (UniqueName: \"kubernetes.io/projected/07b1ae10-2353-4d92-850b-2370581c674f-kube-api-access-s4qmx\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625420 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-service-ca\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625440 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4ngn\" (UniqueName: \"kubernetes.io/projected/b364e7c6-bd4d-4d58-b348-c700933f4d47-kube-api-access-l4ngn\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625455 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9d1e5baf-648e-4327-afc5-7a2835d7fb21-audit-policies\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625470 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh94w\" (UniqueName: \"kubernetes.io/projected/9d1e5baf-648e-4327-afc5-7a2835d7fb21-kube-api-access-hh94w\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625485 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9bxm\" (UniqueName: \"kubernetes.io/projected/7445c312-10f0-4296-ba35-2320e22382e8-kube-api-access-l9bxm\") pod \"downloads-7954f5f757-9dxf2\" (UID: \"7445c312-10f0-4296-ba35-2320e22382e8\") " pod="openshift-console/downloads-7954f5f757-9dxf2" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625500 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-console-oauth-config\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625516 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5xz2\" (UniqueName: \"kubernetes.io/projected/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-kube-api-access-l5xz2\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625531 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-trusted-ca\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625548 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9d1e5baf-648e-4327-afc5-7a2835d7fb21-etcd-client\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625563 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07b1ae10-2353-4d92-850b-2370581c674f-config\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625577 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/681adfb6-f6a8-4934-ade0-a1499bf95d2f-machine-approver-tls\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625590 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3151332c-78d1-439b-bf22-25e60ed60dd9-etcd-client\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625614 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681adfb6-f6a8-4934-ade0-a1499bf95d2f-config\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.625628 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3151332c-78d1-439b-bf22-25e60ed60dd9-node-pullsecrets\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.628326 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-gdm7n"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.628992 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qww9c"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.629371 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.629644 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.629869 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.630681 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.633246 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.634041 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.634829 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.635573 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.635710 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.636412 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.637312 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.637854 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.638309 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.639901 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gkbbs"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.640607 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2c4n7"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.640776 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.641545 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.641926 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-hxhhw"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.643941 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vqmcg"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.644018 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.644214 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.649610 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2hpxq"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.651083 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.652046 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hsz8c"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.654585 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.654605 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p4p2g"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.656429 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-wvj9s"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.658548 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.658635 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.658895 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.663550 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9dxf2"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.665279 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.666286 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.667287 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.668582 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-84p7l"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.669651 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f8tsr"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.670733 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.671758 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.673099 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fzvcr"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.675021 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5dlk4"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.676752 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qww9c"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.678096 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.678365 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.679523 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.680575 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6vpzc"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.682909 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.688300 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hxhhw"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.688339 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.688354 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-pr9hq"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.689758 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pr9hq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.692820 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-d7z6d"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.694416 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.697462 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.699327 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.700573 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.701612 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s2td2"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.702652 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pr9hq"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.704164 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.706108 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gkbbs"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.706296 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2c4n7"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.707943 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7"] Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.718486 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726310 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d1e5baf-648e-4327-afc5-7a2835d7fb21-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726350 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/07b1ae10-2353-4d92-850b-2370581c674f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726379 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-console-config\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726402 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-trusted-ca-bundle\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726426 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726447 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9d1e5baf-648e-4327-afc5-7a2835d7fb21-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726470 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-config\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726493 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726524 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726551 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726575 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07b1ae10-2353-4d92-850b-2370581c674f-serving-cert\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726596 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8dqx\" (UniqueName: \"kubernetes.io/projected/54f4b8f7-8d9c-4bef-af7f-cb2335639e50-kube-api-access-d8dqx\") pod \"cluster-samples-operator-665b6dd947-f5nxk\" (UID: \"54f4b8f7-8d9c-4bef-af7f-cb2335639e50\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726620 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/54f4b8f7-8d9c-4bef-af7f-cb2335639e50-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-f5nxk\" (UID: \"54f4b8f7-8d9c-4bef-af7f-cb2335639e50\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726647 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/45811a4b-98c3-4f54-b772-f1adc68d88c7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rsbdr\" (UID: \"45811a4b-98c3-4f54-b772-f1adc68d88c7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726671 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smqwb\" (UniqueName: \"kubernetes.io/projected/ec936935-d216-4775-be35-f45ed5c5ceca-kube-api-access-smqwb\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726697 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726721 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w26t\" (UniqueName: \"kubernetes.io/projected/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-kube-api-access-5w26t\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726744 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x2px\" (UniqueName: \"kubernetes.io/projected/fada681c-6ade-49e4-b324-1985853bc61e-kube-api-access-8x2px\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726770 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/32a6e5f5-b72c-4e07-a87c-72d7420a1599-proxy-tls\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726791 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68k4h\" (UniqueName: \"kubernetes.io/projected/45811a4b-98c3-4f54-b772-f1adc68d88c7-kube-api-access-68k4h\") pod \"control-plane-machine-set-operator-78cbb6b69f-rsbdr\" (UID: \"45811a4b-98c3-4f54-b772-f1adc68d88c7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726813 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8385c8d3-791b-48c6-9173-84b036a6f294-metrics-tls\") pod \"dns-operator-744455d44c-gkbbs\" (UID: \"8385c8d3-791b-48c6-9173-84b036a6f294\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726839 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ffaf251e-ab8f-4339-905d-5bd8f700beee-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726865 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726907 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726929 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3151332c-78d1-439b-bf22-25e60ed60dd9-encryption-config\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726954 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7zlq\" (UniqueName: \"kubernetes.io/projected/55263751-50ce-4c6b-a3e9-25c7fee34e37-kube-api-access-v7zlq\") pod \"openshift-apiserver-operator-796bbdcf4f-kxgml\" (UID: \"55263751-50ce-4c6b-a3e9-25c7fee34e37\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726977 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jrcf\" (UniqueName: \"kubernetes.io/projected/681adfb6-f6a8-4934-ade0-a1499bf95d2f-kube-api-access-9jrcf\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.726998 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74cbb6cf-4717-439c-b712-67718c154742-serving-cert\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727001 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d1e5baf-648e-4327-afc5-7a2835d7fb21-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727018 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d1e5baf-648e-4327-afc5-7a2835d7fb21-serving-cert\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727040 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727062 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55263751-50ce-4c6b-a3e9-25c7fee34e37-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kxgml\" (UID: \"55263751-50ce-4c6b-a3e9-25c7fee34e37\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727085 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b1c07b60-6159-4cbd-a62c-d684bd8dbd22-signing-cabundle\") pod \"service-ca-9c57cc56f-fzvcr\" (UID: \"b1c07b60-6159-4cbd-a62c-d684bd8dbd22\") " pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727109 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnsfq\" (UniqueName: \"kubernetes.io/projected/b1c07b60-6159-4cbd-a62c-d684bd8dbd22-kube-api-access-wnsfq\") pod \"service-ca-9c57cc56f-fzvcr\" (UID: \"b1c07b60-6159-4cbd-a62c-d684bd8dbd22\") " pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727131 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-default-certificate\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727158 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727182 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffaf251e-ab8f-4339-905d-5bd8f700beee-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727203 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-oauth-serving-cert\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727225 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-audit\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727248 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4dd52d7-4597-4159-a728-012cd2180b35-serving-cert\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727271 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727292 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727313 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-config\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727327 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9d1e5baf-648e-4327-afc5-7a2835d7fb21-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727336 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/32a6e5f5-b72c-4e07-a87c-72d7420a1599-images\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727390 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ffaf251e-ab8f-4339-905d-5bd8f700beee-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727420 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4qmx\" (UniqueName: \"kubernetes.io/projected/07b1ae10-2353-4d92-850b-2370581c674f-kube-api-access-s4qmx\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727448 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-service-ca\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727471 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727497 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v6b2\" (UniqueName: \"kubernetes.io/projected/32a6e5f5-b72c-4e07-a87c-72d7420a1599-kube-api-access-5v6b2\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727522 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9d1e5baf-648e-4327-afc5-7a2835d7fb21-audit-policies\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727531 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/07b1ae10-2353-4d92-850b-2370581c674f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727546 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh94w\" (UniqueName: \"kubernetes.io/projected/9d1e5baf-648e-4327-afc5-7a2835d7fb21-kube-api-access-hh94w\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727568 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9bxm\" (UniqueName: \"kubernetes.io/projected/7445c312-10f0-4296-ba35-2320e22382e8-kube-api-access-l9bxm\") pod \"downloads-7954f5f757-9dxf2\" (UID: \"7445c312-10f0-4296-ba35-2320e22382e8\") " pod="openshift-console/downloads-7954f5f757-9dxf2" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727590 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-console-oauth-config\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727614 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5xz2\" (UniqueName: \"kubernetes.io/projected/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-kube-api-access-l5xz2\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727622 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-console-config\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727636 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-trusted-ca\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727704 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4ngn\" (UniqueName: \"kubernetes.io/projected/b364e7c6-bd4d-4d58-b348-c700933f4d47-kube-api-access-l4ngn\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727727 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-trusted-ca-bundle\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728263 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728309 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07b1ae10-2353-4d92-850b-2370581c674f-config\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728318 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-config\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.727734 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07b1ae10-2353-4d92-850b-2370581c674f-config\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728363 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/681adfb6-f6a8-4934-ade0-a1499bf95d2f-machine-approver-tls\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728390 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3151332c-78d1-439b-bf22-25e60ed60dd9-etcd-client\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728420 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9d1e5baf-648e-4327-afc5-7a2835d7fb21-etcd-client\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728460 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681adfb6-f6a8-4934-ade0-a1499bf95d2f-config\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728486 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3151332c-78d1-439b-bf22-25e60ed60dd9-node-pullsecrets\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728520 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32a6e5f5-b72c-4e07-a87c-72d7420a1599-auth-proxy-config\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728549 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-image-import-ca\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728573 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3151332c-78d1-439b-bf22-25e60ed60dd9-serving-cert\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728596 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b1c07b60-6159-4cbd-a62c-d684bd8dbd22-signing-key\") pod \"service-ca-9c57cc56f-fzvcr\" (UID: \"b1c07b60-6159-4cbd-a62c-d684bd8dbd22\") " pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728623 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhps7\" (UniqueName: \"kubernetes.io/projected/8385c8d3-791b-48c6-9173-84b036a6f294-kube-api-access-hhps7\") pod \"dns-operator-744455d44c-gkbbs\" (UID: \"8385c8d3-791b-48c6-9173-84b036a6f294\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728636 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9d1e5baf-648e-4327-afc5-7a2835d7fb21-audit-policies\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728649 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728674 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3151332c-78d1-439b-bf22-25e60ed60dd9-audit-dir\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728700 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728722 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-metrics-certs\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728745 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl8hb\" (UniqueName: \"kubernetes.io/projected/ffaf251e-ab8f-4339-905d-5bd8f700beee-kube-api-access-fl8hb\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728766 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55263751-50ce-4c6b-a3e9-25c7fee34e37-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kxgml\" (UID: \"55263751-50ce-4c6b-a3e9-25c7fee34e37\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728799 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728860 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-etcd-serving-ca\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728907 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728928 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-config\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728950 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-dir\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728970 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrfln\" (UniqueName: \"kubernetes.io/projected/3151332c-78d1-439b-bf22-25e60ed60dd9-kube-api-access-zrfln\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.728997 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q6np\" (UniqueName: \"kubernetes.io/projected/b4dd52d7-4597-4159-a728-012cd2180b35-kube-api-access-9q6np\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729022 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec936935-d216-4775-be35-f45ed5c5ceca-service-ca-bundle\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729049 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9d1e5baf-648e-4327-afc5-7a2835d7fb21-encryption-config\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729073 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-client-ca\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729093 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-console-serving-cert\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729114 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fada681c-6ade-49e4-b324-1985853bc61e-serving-cert\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729138 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/681adfb6-f6a8-4934-ade0-a1499bf95d2f-auth-proxy-config\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729160 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/07b1ae10-2353-4d92-850b-2370581c674f-service-ca-bundle\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729182 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srckn\" (UniqueName: \"kubernetes.io/projected/74cbb6cf-4717-439c-b712-67718c154742-kube-api-access-srckn\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729204 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729227 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-config\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729251 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-stats-auth\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729277 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9d1e5baf-648e-4327-afc5-7a2835d7fb21-audit-dir\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729306 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-client-ca\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.729744 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.730145 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-etcd-serving-ca\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.730384 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-client-ca\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.730835 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-service-ca\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.731958 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9d1e5baf-648e-4327-afc5-7a2835d7fb21-audit-dir\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.731950 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-dir\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.732135 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/681adfb6-f6a8-4934-ade0-a1499bf95d2f-auth-proxy-config\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.732538 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-config\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.732574 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/07b1ae10-2353-4d92-850b-2370581c674f-service-ca-bundle\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.732955 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3151332c-78d1-439b-bf22-25e60ed60dd9-audit-dir\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.733398 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.733511 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55263751-50ce-4c6b-a3e9-25c7fee34e37-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kxgml\" (UID: \"55263751-50ce-4c6b-a3e9-25c7fee34e37\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.733802 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ffaf251e-ab8f-4339-905d-5bd8f700beee-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.733838 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3151332c-78d1-439b-bf22-25e60ed60dd9-encryption-config\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.734515 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-oauth-serving-cert\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.734602 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.734644 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.734810 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3151332c-78d1-439b-bf22-25e60ed60dd9-node-pullsecrets\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.734966 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-audit\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.735048 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74cbb6cf-4717-439c-b712-67718c154742-serving-cert\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.735589 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681adfb6-f6a8-4934-ade0-a1499bf95d2f-config\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.735723 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-image-import-ca\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.736004 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-console-oauth-config\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.736324 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9d1e5baf-648e-4327-afc5-7a2835d7fb21-encryption-config\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.736322 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/681adfb6-f6a8-4934-ade0-a1499bf95d2f-machine-approver-tls\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.737019 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.737109 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3151332c-78d1-439b-bf22-25e60ed60dd9-etcd-client\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.737556 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.737726 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07b1ae10-2353-4d92-850b-2370581c674f-serving-cert\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.737779 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9d1e5baf-648e-4327-afc5-7a2835d7fb21-etcd-client\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.738068 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.738103 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ffaf251e-ab8f-4339-905d-5bd8f700beee-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.738369 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-console-serving-cert\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.738518 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.738791 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3151332c-78d1-439b-bf22-25e60ed60dd9-serving-cert\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.739559 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3151332c-78d1-439b-bf22-25e60ed60dd9-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.739648 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55263751-50ce-4c6b-a3e9-25c7fee34e37-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kxgml\" (UID: \"55263751-50ce-4c6b-a3e9-25c7fee34e37\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.740028 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/54f4b8f7-8d9c-4bef-af7f-cb2335639e50-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-f5nxk\" (UID: \"54f4b8f7-8d9c-4bef-af7f-cb2335639e50\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.740716 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d1e5baf-648e-4327-afc5-7a2835d7fb21-serving-cert\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.758553 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.778755 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.798709 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.818229 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.831228 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smqwb\" (UniqueName: \"kubernetes.io/projected/ec936935-d216-4775-be35-f45ed5c5ceca-kube-api-access-smqwb\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.831295 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/45811a4b-98c3-4f54-b772-f1adc68d88c7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rsbdr\" (UID: \"45811a4b-98c3-4f54-b772-f1adc68d88c7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.831346 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68k4h\" (UniqueName: \"kubernetes.io/projected/45811a4b-98c3-4f54-b772-f1adc68d88c7-kube-api-access-68k4h\") pod \"control-plane-machine-set-operator-78cbb6b69f-rsbdr\" (UID: \"45811a4b-98c3-4f54-b772-f1adc68d88c7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.831406 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8385c8d3-791b-48c6-9173-84b036a6f294-metrics-tls\") pod \"dns-operator-744455d44c-gkbbs\" (UID: \"8385c8d3-791b-48c6-9173-84b036a6f294\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.831475 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/32a6e5f5-b72c-4e07-a87c-72d7420a1599-proxy-tls\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.831537 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b1c07b60-6159-4cbd-a62c-d684bd8dbd22-signing-cabundle\") pod \"service-ca-9c57cc56f-fzvcr\" (UID: \"b1c07b60-6159-4cbd-a62c-d684bd8dbd22\") " pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.831628 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnsfq\" (UniqueName: \"kubernetes.io/projected/b1c07b60-6159-4cbd-a62c-d684bd8dbd22-kube-api-access-wnsfq\") pod \"service-ca-9c57cc56f-fzvcr\" (UID: \"b1c07b60-6159-4cbd-a62c-d684bd8dbd22\") " pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.831721 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-default-certificate\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.831944 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/32a6e5f5-b72c-4e07-a87c-72d7420a1599-images\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.832080 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v6b2\" (UniqueName: \"kubernetes.io/projected/32a6e5f5-b72c-4e07-a87c-72d7420a1599-kube-api-access-5v6b2\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.832203 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32a6e5f5-b72c-4e07-a87c-72d7420a1599-auth-proxy-config\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.832238 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b1c07b60-6159-4cbd-a62c-d684bd8dbd22-signing-key\") pod \"service-ca-9c57cc56f-fzvcr\" (UID: \"b1c07b60-6159-4cbd-a62c-d684bd8dbd22\") " pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.832269 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhps7\" (UniqueName: \"kubernetes.io/projected/8385c8d3-791b-48c6-9173-84b036a6f294-kube-api-access-hhps7\") pod \"dns-operator-744455d44c-gkbbs\" (UID: \"8385c8d3-791b-48c6-9173-84b036a6f294\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.832326 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-metrics-certs\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.832390 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec936935-d216-4775-be35-f45ed5c5ceca-service-ca-bundle\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.832482 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-stats-auth\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.832949 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32a6e5f5-b72c-4e07-a87c-72d7420a1599-auth-proxy-config\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.838749 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.858916 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.879371 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.899275 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.918413 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.938916 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.946821 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/45811a4b-98c3-4f54-b772-f1adc68d88c7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rsbdr\" (UID: \"45811a4b-98c3-4f54-b772-f1adc68d88c7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.958579 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 21:53:31 crc kubenswrapper[4931]: I1125 21:53:31.978167 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.008041 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.019293 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.039765 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.059105 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.078142 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.099234 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.118550 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.138802 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.158452 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.179385 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.198568 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.218620 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.238377 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.258964 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.279643 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.299074 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.318528 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.342610 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.358577 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.379543 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.401533 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.419433 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.449288 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.458688 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.478729 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.498998 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.521457 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.538983 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.558768 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.579163 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.592128 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b1c07b60-6159-4cbd-a62c-d684bd8dbd22-signing-key\") pod \"service-ca-9c57cc56f-fzvcr\" (UID: \"b1c07b60-6159-4cbd-a62c-d684bd8dbd22\") " pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.599168 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.603863 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b1c07b60-6159-4cbd-a62c-d684bd8dbd22-signing-cabundle\") pod \"service-ca-9c57cc56f-fzvcr\" (UID: \"b1c07b60-6159-4cbd-a62c-d684bd8dbd22\") " pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.618988 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.637350 4931 request.go:700] Waited for 1.015783136s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/secrets?fieldSelector=metadata.name%3Dmultus-admission-controller-secret&limit=500&resourceVersion=0 Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.639379 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.659750 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.679313 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.699940 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.719780 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.728222 4931 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-idp-0-file-data: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.728333 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data podName:b364e7c6-bd4d-4d58-b348-c700933f4d47 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.22830299 +0000 UTC m=+151.584815149 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-idp-0-file-data" (UniqueName: "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data") pod "oauth-openshift-558db77b4-p4p2g" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.728228 4931 configmap.go:193] Couldn't get configMap openshift-console-operator/trusted-ca: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.728672 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-trusted-ca podName:b4dd52d7-4597-4159-a728-012cd2180b35 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.228654019 +0000 UTC m=+151.585166168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca" (UniqueName: "kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-trusted-ca") pod "console-operator-58897d9998-vqmcg" (UID: "b4dd52d7-4597-4159-a728-012cd2180b35") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.731482 4931 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.731568 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert podName:b364e7c6-bd4d-4d58-b348-c700933f4d47 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.231547048 +0000 UTC m=+151.588059207 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-p4p2g" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.732642 4931 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.732716 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fada681c-6ade-49e4-b324-1985853bc61e-serving-cert podName:fada681c-6ade-49e4-b324-1985853bc61e nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.2326984 +0000 UTC m=+151.589210549 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/fada681c-6ade-49e4-b324-1985853bc61e-serving-cert") pod "route-controller-manager-6576b87f9c-454st" (UID: "fada681c-6ade-49e4-b324-1985853bc61e") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734082 4931 configmap.go:193] Couldn't get configMap openshift-controller-manager-operator/openshift-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734117 4931 configmap.go:193] Couldn't get configMap openshift-console-operator/console-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734195 4931 secret.go:188] Couldn't get secret openshift-console-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734222 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-config podName:0211b0cb-4e13-45ce-aab7-a0e6848c2bef nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.234189221 +0000 UTC m=+151.590701380 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-config") pod "openshift-controller-manager-operator-756b6f6bc6-rwnnf" (UID: "0211b0cb-4e13-45ce-aab7-a0e6848c2bef") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734222 4931 secret.go:188] Couldn't get secret openshift-controller-manager-operator/openshift-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734269 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-config podName:b4dd52d7-4597-4159-a728-012cd2180b35 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.234245833 +0000 UTC m=+151.590758042 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-config") pod "console-operator-58897d9998-vqmcg" (UID: "b4dd52d7-4597-4159-a728-012cd2180b35") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734219 4931 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734326 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-serving-cert podName:0211b0cb-4e13-45ce-aab7-a0e6848c2bef nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.234298614 +0000 UTC m=+151.590810783 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-serving-cert") pod "openshift-controller-manager-operator-756b6f6bc6-rwnnf" (UID: "0211b0cb-4e13-45ce-aab7-a0e6848c2bef") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734355 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4dd52d7-4597-4159-a728-012cd2180b35-serving-cert podName:b4dd52d7-4597-4159-a728-012cd2180b35 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.234340995 +0000 UTC m=+151.590853144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/b4dd52d7-4597-4159-a728-012cd2180b35-serving-cert") pod "console-operator-58897d9998-vqmcg" (UID: "b4dd52d7-4597-4159-a728-012cd2180b35") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.734374 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-client-ca podName:fada681c-6ade-49e4-b324-1985853bc61e nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.234364796 +0000 UTC m=+151.590876955 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-client-ca") pod "route-controller-manager-6576b87f9c-454st" (UID: "fada681c-6ade-49e4-b324-1985853bc61e") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.735285 4931 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.735316 4931 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-session: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.735341 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-config podName:fada681c-6ade-49e4-b324-1985853bc61e nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.235326573 +0000 UTC m=+151.591838722 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-config") pod "route-controller-manager-6576b87f9c-454st" (UID: "fada681c-6ade-49e4-b324-1985853bc61e") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.735344 4931 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.735406 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-session podName:b364e7c6-bd4d-4d58-b348-c700933f4d47 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.235377484 +0000 UTC m=+151.591889643 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-session") pod "oauth-openshift-558db77b4-p4p2g" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.735441 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies podName:b364e7c6-bd4d-4d58-b348-c700933f4d47 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.235425055 +0000 UTC m=+151.591937294 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies") pod "oauth-openshift-558db77b4-p4p2g" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.738836 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.758661 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.779311 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.800494 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.809836 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-metrics-certs\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.818828 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.832764 4931 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.832787 4931 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.832835 4931 secret.go:188] Couldn't get secret openshift-ingress/router-stats-default: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.832860 4931 secret.go:188] Couldn't get secret openshift-ingress/router-certs-default: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.832934 4931 secret.go:188] Couldn't get secret openshift-dns-operator/metrics-tls: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.832904 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/32a6e5f5-b72c-4e07-a87c-72d7420a1599-images podName:32a6e5f5-b72c-4e07-a87c-72d7420a1599 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.332849415 +0000 UTC m=+151.689361564 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/32a6e5f5-b72c-4e07-a87c-72d7420a1599-images") pod "machine-config-operator-74547568cd-w6tf9" (UID: "32a6e5f5-b72c-4e07-a87c-72d7420a1599") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.833011 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32a6e5f5-b72c-4e07-a87c-72d7420a1599-proxy-tls podName:32a6e5f5-b72c-4e07-a87c-72d7420a1599 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.332987459 +0000 UTC m=+151.689499658 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/32a6e5f5-b72c-4e07-a87c-72d7420a1599-proxy-tls") pod "machine-config-operator-74547568cd-w6tf9" (UID: "32a6e5f5-b72c-4e07-a87c-72d7420a1599") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.833033 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-stats-auth podName:ec936935-d216-4775-be35-f45ed5c5ceca nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.33302528 +0000 UTC m=+151.689537509 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "stats-auth" (UniqueName: "kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-stats-auth") pod "router-default-5444994796-gdm7n" (UID: "ec936935-d216-4775-be35-f45ed5c5ceca") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.833055 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-default-certificate podName:ec936935-d216-4775-be35-f45ed5c5ceca nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.33304419 +0000 UTC m=+151.689556529 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-certificate" (UniqueName: "kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-default-certificate") pod "router-default-5444994796-gdm7n" (UID: "ec936935-d216-4775-be35-f45ed5c5ceca") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.833078 4931 configmap.go:193] Couldn't get configMap openshift-ingress/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.833091 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8385c8d3-791b-48c6-9173-84b036a6f294-metrics-tls podName:8385c8d3-791b-48c6-9173-84b036a6f294 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.333078341 +0000 UTC m=+151.689590590 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/8385c8d3-791b-48c6-9173-84b036a6f294-metrics-tls") pod "dns-operator-744455d44c-gkbbs" (UID: "8385c8d3-791b-48c6-9173-84b036a6f294") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: E1125 21:53:32.833128 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ec936935-d216-4775-be35-f45ed5c5ceca-service-ca-bundle podName:ec936935-d216-4775-be35-f45ed5c5ceca nodeName:}" failed. No retries permitted until 2025-11-25 21:53:33.333101882 +0000 UTC m=+151.689614001 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/ec936935-d216-4775-be35-f45ed5c5ceca-service-ca-bundle") pod "router-default-5444994796-gdm7n" (UID: "ec936935-d216-4775-be35-f45ed5c5ceca") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.837899 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.859086 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.878512 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.898810 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.939441 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.958856 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.979936 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 21:53:32 crc kubenswrapper[4931]: I1125 21:53:32.999688 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.019718 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.039474 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.059061 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.078708 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.098739 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.119697 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.140737 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.158486 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.179859 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.199047 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.218785 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.238969 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.254661 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.254719 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4dd52d7-4597-4159-a728-012cd2180b35-serving-cert\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.254763 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.254797 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.254829 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-config\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.254976 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-trusted-ca\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.255065 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.255099 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-config\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.255172 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-client-ca\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.255220 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fada681c-6ade-49e4-b324-1985853bc61e-serving-cert\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.255251 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.255313 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.259503 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.278941 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.298755 4931 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.319557 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.338827 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.356945 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/32a6e5f5-b72c-4e07-a87c-72d7420a1599-proxy-tls\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.357008 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8385c8d3-791b-48c6-9173-84b036a6f294-metrics-tls\") pod \"dns-operator-744455d44c-gkbbs\" (UID: \"8385c8d3-791b-48c6-9173-84b036a6f294\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.357102 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-default-certificate\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.357173 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/32a6e5f5-b72c-4e07-a87c-72d7420a1599-images\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.357328 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec936935-d216-4775-be35-f45ed5c5ceca-service-ca-bundle\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.357397 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-stats-auth\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.358519 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.358761 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/32a6e5f5-b72c-4e07-a87c-72d7420a1599-images\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.359429 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec936935-d216-4775-be35-f45ed5c5ceca-service-ca-bundle\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.361510 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-default-certificate\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.363522 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8385c8d3-791b-48c6-9173-84b036a6f294-metrics-tls\") pod \"dns-operator-744455d44c-gkbbs\" (UID: \"8385c8d3-791b-48c6-9173-84b036a6f294\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.364591 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/32a6e5f5-b72c-4e07-a87c-72d7420a1599-proxy-tls\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.368952 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ec936935-d216-4775-be35-f45ed5c5ceca-stats-auth\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.380312 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.418476 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.438466 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.459156 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.478725 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.499396 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.518902 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.538975 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.606561 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ffaf251e-ab8f-4339-905d-5bd8f700beee-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.641475 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh94w\" (UniqueName: \"kubernetes.io/projected/9d1e5baf-648e-4327-afc5-7a2835d7fb21-kube-api-access-hh94w\") pod \"apiserver-7bbb656c7d-8jz7m\" (UID: \"9d1e5baf-648e-4327-afc5-7a2835d7fb21\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.657283 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9bxm\" (UniqueName: \"kubernetes.io/projected/7445c312-10f0-4296-ba35-2320e22382e8-kube-api-access-l9bxm\") pod \"downloads-7954f5f757-9dxf2\" (UID: \"7445c312-10f0-4296-ba35-2320e22382e8\") " pod="openshift-console/downloads-7954f5f757-9dxf2" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.657379 4931 request.go:700] Waited for 1.928597758s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication-operator/serviceaccounts/authentication-operator/token Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.681535 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4qmx\" (UniqueName: \"kubernetes.io/projected/07b1ae10-2353-4d92-850b-2370581c674f-kube-api-access-s4qmx\") pod \"authentication-operator-69f744f599-wvj9s\" (UID: \"07b1ae10-2353-4d92-850b-2370581c674f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.696942 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5xz2\" (UniqueName: \"kubernetes.io/projected/69dfd797-b979-42a6-a4d5-d95e48bf7aa2-kube-api-access-l5xz2\") pod \"console-f9d7485db-2hpxq\" (UID: \"69dfd797-b979-42a6-a4d5-d95e48bf7aa2\") " pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.717447 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7zlq\" (UniqueName: \"kubernetes.io/projected/55263751-50ce-4c6b-a3e9-25c7fee34e37-kube-api-access-v7zlq\") pod \"openshift-apiserver-operator-796bbdcf4f-kxgml\" (UID: \"55263751-50ce-4c6b-a3e9-25c7fee34e37\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.725564 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.735072 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jrcf\" (UniqueName: \"kubernetes.io/projected/681adfb6-f6a8-4934-ade0-a1499bf95d2f-kube-api-access-9jrcf\") pod \"machine-approver-56656f9798-g78xj\" (UID: \"681adfb6-f6a8-4934-ade0-a1499bf95d2f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.739368 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.747759 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.754122 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrfln\" (UniqueName: \"kubernetes.io/projected/3151332c-78d1-439b-bf22-25e60ed60dd9-kube-api-access-zrfln\") pod \"apiserver-76f77b778f-f8tsr\" (UID: \"3151332c-78d1-439b-bf22-25e60ed60dd9\") " pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:33 crc kubenswrapper[4931]: W1125 21:53:33.761427 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod681adfb6_f6a8_4934_ade0_a1499bf95d2f.slice/crio-5840b0b3e347e4f4c1367a0ab8a700d59d1fd7241cb3c2bd7cf99c151527a57e WatchSource:0}: Error finding container 5840b0b3e347e4f4c1367a0ab8a700d59d1fd7241cb3c2bd7cf99c151527a57e: Status 404 returned error can't find the container with id 5840b0b3e347e4f4c1367a0ab8a700d59d1fd7241cb3c2bd7cf99c151527a57e Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.764453 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.771755 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.775654 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8dqx\" (UniqueName: \"kubernetes.io/projected/54f4b8f7-8d9c-4bef-af7f-cb2335639e50-kube-api-access-d8dqx\") pod \"cluster-samples-operator-665b6dd947-f5nxk\" (UID: \"54f4b8f7-8d9c-4bef-af7f-cb2335639e50\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.780345 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9dxf2" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.815387 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srckn\" (UniqueName: \"kubernetes.io/projected/74cbb6cf-4717-439c-b712-67718c154742-kube-api-access-srckn\") pod \"controller-manager-879f6c89f-kvz7t\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.861315 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl8hb\" (UniqueName: \"kubernetes.io/projected/ffaf251e-ab8f-4339-905d-5bd8f700beee-kube-api-access-fl8hb\") pod \"cluster-image-registry-operator-dc59b4c8b-tqm7k\" (UID: \"ffaf251e-ab8f-4339-905d-5bd8f700beee\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.876547 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smqwb\" (UniqueName: \"kubernetes.io/projected/ec936935-d216-4775-be35-f45ed5c5ceca-kube-api-access-smqwb\") pod \"router-default-5444994796-gdm7n\" (UID: \"ec936935-d216-4775-be35-f45ed5c5ceca\") " pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.882850 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.894161 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68k4h\" (UniqueName: \"kubernetes.io/projected/45811a4b-98c3-4f54-b772-f1adc68d88c7-kube-api-access-68k4h\") pod \"control-plane-machine-set-operator-78cbb6b69f-rsbdr\" (UID: \"45811a4b-98c3-4f54-b772-f1adc68d88c7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.911510 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnsfq\" (UniqueName: \"kubernetes.io/projected/b1c07b60-6159-4cbd-a62c-d684bd8dbd22-kube-api-access-wnsfq\") pod \"service-ca-9c57cc56f-fzvcr\" (UID: \"b1c07b60-6159-4cbd-a62c-d684bd8dbd22\") " pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.924491 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.942300 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v6b2\" (UniqueName: \"kubernetes.io/projected/32a6e5f5-b72c-4e07-a87c-72d7420a1599-kube-api-access-5v6b2\") pod \"machine-config-operator-74547568cd-w6tf9\" (UID: \"32a6e5f5-b72c-4e07-a87c-72d7420a1599\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.950746 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.951954 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.952844 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhps7\" (UniqueName: \"kubernetes.io/projected/8385c8d3-791b-48c6-9173-84b036a6f294-kube-api-access-hhps7\") pod \"dns-operator-744455d44c-gkbbs\" (UID: \"8385c8d3-791b-48c6-9173-84b036a6f294\") " pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.959039 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.967859 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.977971 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-trusted-ca\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.979459 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.990525 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.997079 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" Nov 25 21:53:33 crc kubenswrapper[4931]: I1125 21:53:33.999608 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.014431 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.019240 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.042647 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.055404 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.059015 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4dd52d7-4597-4159-a728-012cd2180b35-serving-cert\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.059048 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.080663 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.094446 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.100080 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 21:53:34 crc kubenswrapper[4931]: W1125 21:53:34.117300 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d1e5baf_648e_4327_afc5_7a2835d7fb21.slice/crio-b9bc302d69261cbfe9a1dd1c363e71d7849fa757bc7e56767170cdfe453cefa2 WatchSource:0}: Error finding container b9bc302d69261cbfe9a1dd1c363e71d7849fa757bc7e56767170cdfe453cefa2: Status 404 returned error can't find the container with id b9bc302d69261cbfe9a1dd1c363e71d7849fa757bc7e56767170cdfe453cefa2 Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.119434 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.133243 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x2px\" (UniqueName: \"kubernetes.io/projected/fada681c-6ade-49e4-b324-1985853bc61e-kube-api-access-8x2px\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.138440 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.141619 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-wvj9s"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.146775 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.161796 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.171410 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.172751 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fada681c-6ade-49e4-b324-1985853bc61e-serving-cert\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.178550 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.180473 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.187934 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-config\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:34 crc kubenswrapper[4931]: W1125 21:53:34.196316 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55263751_50ce_4c6b_a3e9_25c7fee34e37.slice/crio-ab4b440cd0735bfe81ae0ed830654957abd832332de4987229572065bce7df79 WatchSource:0}: Error finding container ab4b440cd0735bfe81ae0ed830654957abd832332de4987229572065bce7df79: Status 404 returned error can't find the container with id ab4b440cd0735bfe81ae0ed830654957abd832332de4987229572065bce7df79 Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.198626 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.206543 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-client-ca\") pod \"route-controller-manager-6576b87f9c-454st\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.218318 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.226394 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4ngn\" (UniqueName: \"kubernetes.io/projected/b364e7c6-bd4d-4d58-b348-c700933f4d47-kube-api-access-l4ngn\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.239097 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.248618 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4dd52d7-4597-4159-a728-012cd2180b35-config\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.257075 4931 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.257169 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert podName:b364e7c6-bd4d-4d58-b348-c700933f4d47 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.257148385 +0000 UTC m=+153.613660504 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-p4p2g" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.257700 4931 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-idp-0-file-data: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.257727 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data podName:b364e7c6-bd4d-4d58-b348-c700933f4d47 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.25772005 +0000 UTC m=+153.614232169 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-idp-0-file-data" (UniqueName: "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data") pod "oauth-openshift-558db77b4-p4p2g" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.257740 4931 secret.go:188] Couldn't get secret openshift-controller-manager-operator/openshift-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.257764 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-serving-cert podName:0211b0cb-4e13-45ce-aab7-a0e6848c2bef nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.257756691 +0000 UTC m=+153.614268810 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-serving-cert") pod "openshift-controller-manager-operator-756b6f6bc6-rwnnf" (UID: "0211b0cb-4e13-45ce-aab7-a0e6848c2bef") : failed to sync secret cache: timed out waiting for the condition Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.258437 4931 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.258556 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies podName:b364e7c6-bd4d-4d58-b348-c700933f4d47 nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.258534863 +0000 UTC m=+153.615047012 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies") pod "oauth-openshift-558db77b4-p4p2g" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47") : failed to sync configmap cache: timed out waiting for the condition Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.259814 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.266763 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2hpxq"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.271149 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q6np\" (UniqueName: \"kubernetes.io/projected/b4dd52d7-4597-4159-a728-012cd2180b35-kube-api-access-9q6np\") pod \"console-operator-58897d9998-vqmcg\" (UID: \"b4dd52d7-4597-4159-a728-012cd2180b35\") " pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.272287 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9dxf2"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.278483 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.298328 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.298944 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f8tsr"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.302145 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.307694 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gkbbs"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.318763 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.340065 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.349689 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" event={"ID":"3151332c-78d1-439b-bf22-25e60ed60dd9","Type":"ContainerStarted","Data":"43b00391f1004f4a1d8d7030d99f1668baadd746095919aae4faa5538af70adb"} Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.350413 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" event={"ID":"9d1e5baf-648e-4327-afc5-7a2835d7fb21","Type":"ContainerStarted","Data":"b9bc302d69261cbfe9a1dd1c363e71d7849fa757bc7e56767170cdfe453cefa2"} Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.351050 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" event={"ID":"07b1ae10-2353-4d92-850b-2370581c674f","Type":"ContainerStarted","Data":"15e46bd30d43ea0487ae1316a1477d2f67f6951dcf9b149f122ba45fb683ee74"} Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.353822 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" event={"ID":"8385c8d3-791b-48c6-9173-84b036a6f294","Type":"ContainerStarted","Data":"e6945a4d84314a49ed7dd987348fc75b340f9ea33d45bff0e0772596a1265d87"} Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.357289 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2hpxq" event={"ID":"69dfd797-b979-42a6-a4d5-d95e48bf7aa2","Type":"ContainerStarted","Data":"c4aa5004381d3a562d530badad8968f73ca64c73bebd75914fa39a2aa6c8bdb7"} Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.358047 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" event={"ID":"55263751-50ce-4c6b-a3e9-25c7fee34e37","Type":"ContainerStarted","Data":"ab4b440cd0735bfe81ae0ed830654957abd832332de4987229572065bce7df79"} Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.358357 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.358992 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gdm7n" event={"ID":"ec936935-d216-4775-be35-f45ed5c5ceca","Type":"ContainerStarted","Data":"5020b17886d069d47d0c1d28910e43596c60822acd8a5d4412d2a34a5a988846"} Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.360929 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" event={"ID":"681adfb6-f6a8-4934-ade0-a1499bf95d2f","Type":"ContainerStarted","Data":"5840b0b3e347e4f4c1367a0ab8a700d59d1fd7241cb3c2bd7cf99c151527a57e"} Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.364543 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9dxf2" event={"ID":"7445c312-10f0-4296-ba35-2320e22382e8","Type":"ContainerStarted","Data":"792fc1312ab042c814a85b343d4070a30e13687c8c7b8df40046cb4c528bdc61"} Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.379616 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.398750 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.401585 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.402830 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.419313 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.423831 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w26t\" (UniqueName: \"kubernetes.io/projected/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-kube-api-access-5w26t\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.453957 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.456226 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.456262 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kvz7t"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.456283 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk"] Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.456297 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fzvcr"] Nov 25 21:53:34 crc kubenswrapper[4931]: W1125 21:53:34.463534 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffaf251e_ab8f_4339_905d_5bd8f700beee.slice/crio-fdef55b511b3fc1948b077bde93dd00de70e964a21ddb3ff1535340fbfd435c9 WatchSource:0}: Error finding container fdef55b511b3fc1948b077bde93dd00de70e964a21ddb3ff1535340fbfd435c9: Status 404 returned error can't find the container with id fdef55b511b3fc1948b077bde93dd00de70e964a21ddb3ff1535340fbfd435c9 Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.560623 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdnrc\" (UniqueName: \"kubernetes.io/projected/396991bf-e7c0-448e-a2fc-a626f0bd43b6-kube-api-access-fdnrc\") pod \"kube-storage-version-migrator-operator-b67b599dd-vgppb\" (UID: \"396991bf-e7c0-448e-a2fc-a626f0bd43b6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.561226 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-certificates\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.561278 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/396991bf-e7c0-448e-a2fc-a626f0bd43b6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vgppb\" (UID: \"396991bf-e7c0-448e-a2fc-a626f0bd43b6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.561312 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wfjj\" (UniqueName: \"kubernetes.io/projected/bf866554-0f21-4262-9b8e-32d9c0964488-kube-api-access-7wfjj\") pod \"catalog-operator-68c6474976-gkwhd\" (UID: \"bf866554-0f21-4262-9b8e-32d9c0964488\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.563438 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ab99d3e-35d5-4a34-b195-fc53434b5168-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmjcj\" (UID: \"1ab99d3e-35d5-4a34-b195-fc53434b5168\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.563868 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42040525-2c1c-46bf-94e6-b23e05b7b3ae-installation-pull-secrets\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.564037 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.564180 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c35b8c1e-cb44-4e0b-9732-0132cdb4e10a-config\") pod \"service-ca-operator-777779d784-qww9c\" (UID: \"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.564221 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vqmcg"] Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.564676 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.064660121 +0000 UTC m=+153.421172340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.565680 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c35b8c1e-cb44-4e0b-9732-0132cdb4e10a-serving-cert\") pod \"service-ca-operator-777779d784-qww9c\" (UID: \"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.565795 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws246\" (UniqueName: \"kubernetes.io/projected/c35b8c1e-cb44-4e0b-9732-0132cdb4e10a-kube-api-access-ws246\") pod \"service-ca-operator-777779d784-qww9c\" (UID: \"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.567599 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-trusted-ca\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.567654 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53a6c85d-8623-46b6-9d53-860f1ff80f00-webhook-cert\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.567739 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bf866554-0f21-4262-9b8e-32d9c0964488-srv-cert\") pod \"catalog-operator-68c6474976-gkwhd\" (UID: \"bf866554-0f21-4262-9b8e-32d9c0964488\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.567858 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ab99d3e-35d5-4a34-b195-fc53434b5168-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmjcj\" (UID: \"1ab99d3e-35d5-4a34-b195-fc53434b5168\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.567916 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/53a6c85d-8623-46b6-9d53-860f1ff80f00-tmpfs\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.568142 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ab99d3e-35d5-4a34-b195-fc53434b5168-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmjcj\" (UID: \"1ab99d3e-35d5-4a34-b195-fc53434b5168\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.568219 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf6f8\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-kube-api-access-pf6f8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.568390 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-bound-sa-token\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.568434 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bf866554-0f21-4262-9b8e-32d9c0964488-profile-collector-cert\") pod \"catalog-operator-68c6474976-gkwhd\" (UID: \"bf866554-0f21-4262-9b8e-32d9c0964488\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.568465 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t65tp\" (UniqueName: \"kubernetes.io/projected/53a6c85d-8623-46b6-9d53-860f1ff80f00-kube-api-access-t65tp\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.568498 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/396991bf-e7c0-448e-a2fc-a626f0bd43b6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vgppb\" (UID: \"396991bf-e7c0-448e-a2fc-a626f0bd43b6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.568741 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-tls\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.568787 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42040525-2c1c-46bf-94e6-b23e05b7b3ae-ca-trust-extracted\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.570646 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53a6c85d-8623-46b6-9d53-860f1ff80f00-apiservice-cert\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.671615 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.683339 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.183315995 +0000 UTC m=+153.539828114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683425 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-bound-sa-token\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683462 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t65tp\" (UniqueName: \"kubernetes.io/projected/53a6c85d-8623-46b6-9d53-860f1ff80f00-kube-api-access-t65tp\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683510 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpc99\" (UniqueName: \"kubernetes.io/projected/5af8ff72-8043-4f10-8357-10ce2ff5123b-kube-api-access-bpc99\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683536 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42040525-2c1c-46bf-94e6-b23e05b7b3ae-ca-trust-extracted\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683568 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c036f4c8-f544-4f0d-b1c6-51944420430c-config-volume\") pod \"collect-profiles-29401785-v5cgr\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683588 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8sv8\" (UniqueName: \"kubernetes.io/projected/45bb8c11-1f6e-4f8a-a67d-b2f531211ba4-kube-api-access-w8sv8\") pod \"dns-default-hxhhw\" (UID: \"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4\") " pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683623 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-624ws\" (UniqueName: \"kubernetes.io/projected/640d1c70-7a52-4f0d-a4df-a65d4b247acb-kube-api-access-624ws\") pod \"olm-operator-6b444d44fb-hpqqz\" (UID: \"640d1c70-7a52-4f0d-a4df-a65d4b247acb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683648 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5af8ff72-8043-4f10-8357-10ce2ff5123b-trusted-ca\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683670 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-certificates\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683696 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/396991bf-e7c0-448e-a2fc-a626f0bd43b6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vgppb\" (UID: \"396991bf-e7c0-448e-a2fc-a626f0bd43b6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683715 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9fad68f4-2997-42c8-aeb2-937d220f7345-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kxjw7\" (UID: \"9fad68f4-2997-42c8-aeb2-937d220f7345\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683738 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c3c979d-1a74-40a6-b24b-572c0ab97f63-etcd-service-ca\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683761 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6110aafd-c076-4fa4-a0b6-f5f434b9cfea-config\") pod \"kube-apiserver-operator-766d6c64bb-wnjvs\" (UID: \"6110aafd-c076-4fa4-a0b6-f5f434b9cfea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683783 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45bb8c11-1f6e-4f8a-a67d-b2f531211ba4-config-volume\") pod \"dns-default-hxhhw\" (UID: \"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4\") " pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683831 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c3c979d-1a74-40a6-b24b-572c0ab97f63-serving-cert\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683855 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zr9q\" (UniqueName: \"kubernetes.io/projected/842021b8-6f3c-4eab-bcba-0b27f6736e9e-kube-api-access-9zr9q\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683888 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wfjj\" (UniqueName: \"kubernetes.io/projected/bf866554-0f21-4262-9b8e-32d9c0964488-kube-api-access-7wfjj\") pod \"catalog-operator-68c6474976-gkwhd\" (UID: \"bf866554-0f21-4262-9b8e-32d9c0964488\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683905 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-socket-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683924 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9fad68f4-2997-42c8-aeb2-937d220f7345-proxy-tls\") pod \"machine-config-controller-84d6567774-kxjw7\" (UID: \"9fad68f4-2997-42c8-aeb2-937d220f7345\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683951 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-registration-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.683983 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684004 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c35b8c1e-cb44-4e0b-9732-0132cdb4e10a-config\") pod \"service-ca-operator-777779d784-qww9c\" (UID: \"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684027 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/640d1c70-7a52-4f0d-a4df-a65d4b247acb-srv-cert\") pod \"olm-operator-6b444d44fb-hpqqz\" (UID: \"640d1c70-7a52-4f0d-a4df-a65d4b247acb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684054 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f421de15-2d9f-46ef-82da-414130bdf02e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s2td2\" (UID: \"f421de15-2d9f-46ef-82da-414130bdf02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684085 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/35c9fc97-1683-4083-9090-d1cfcc0a605e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bmxc6\" (UID: \"35c9fc97-1683-4083-9090-d1cfcc0a605e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684111 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/842021b8-6f3c-4eab-bcba-0b27f6736e9e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684138 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-trusted-ca\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684159 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53a6c85d-8623-46b6-9d53-860f1ff80f00-webhook-cert\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684183 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7ct7\" (UniqueName: \"kubernetes.io/projected/9fad68f4-2997-42c8-aeb2-937d220f7345-kube-api-access-b7ct7\") pod \"machine-config-controller-84d6567774-kxjw7\" (UID: \"9fad68f4-2997-42c8-aeb2-937d220f7345\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684208 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bf866554-0f21-4262-9b8e-32d9c0964488-srv-cert\") pod \"catalog-operator-68c6474976-gkwhd\" (UID: \"bf866554-0f21-4262-9b8e-32d9c0964488\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684230 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl5fz\" (UniqueName: \"kubernetes.io/projected/8c3c979d-1a74-40a6-b24b-572c0ab97f63-kube-api-access-pl5fz\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684252 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6110aafd-c076-4fa4-a0b6-f5f434b9cfea-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wnjvs\" (UID: \"6110aafd-c076-4fa4-a0b6-f5f434b9cfea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684278 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3477fcf-270d-4b95-a581-3cfe8770cd0d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r4pgh\" (UID: \"f3477fcf-270d-4b95-a581-3cfe8770cd0d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684304 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ab99d3e-35d5-4a34-b195-fc53434b5168-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmjcj\" (UID: \"1ab99d3e-35d5-4a34-b195-fc53434b5168\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684334 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c036f4c8-f544-4f0d-b1c6-51944420430c-secret-volume\") pod \"collect-profiles-29401785-v5cgr\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684359 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5af8ff72-8043-4f10-8357-10ce2ff5123b-metrics-tls\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684386 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ab99d3e-35d5-4a34-b195-fc53434b5168-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmjcj\" (UID: \"1ab99d3e-35d5-4a34-b195-fc53434b5168\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684408 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrtmj\" (UniqueName: \"kubernetes.io/projected/2b04f34b-bb66-4247-92e4-c447085d9144-kube-api-access-jrtmj\") pod \"openshift-config-operator-7777fb866f-9wxs5\" (UID: \"2b04f34b-bb66-4247-92e4-c447085d9144\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684431 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnzgk\" (UniqueName: \"kubernetes.io/projected/35c9fc97-1683-4083-9090-d1cfcc0a605e-kube-api-access-hnzgk\") pod \"package-server-manager-789f6589d5-bmxc6\" (UID: \"35c9fc97-1683-4083-9090-d1cfcc0a605e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684455 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf6f8\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-kube-api-access-pf6f8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684479 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842021b8-6f3c-4eab-bcba-0b27f6736e9e-config\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684516 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bf866554-0f21-4262-9b8e-32d9c0964488-profile-collector-cert\") pod \"catalog-operator-68c6474976-gkwhd\" (UID: \"bf866554-0f21-4262-9b8e-32d9c0964488\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684540 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/396991bf-e7c0-448e-a2fc-a626f0bd43b6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vgppb\" (UID: \"396991bf-e7c0-448e-a2fc-a626f0bd43b6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684560 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8c3c979d-1a74-40a6-b24b-572c0ab97f63-etcd-ca\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684579 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-tls\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.684629 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b04f34b-bb66-4247-92e4-c447085d9144-serving-cert\") pod \"openshift-config-operator-7777fb866f-9wxs5\" (UID: \"2b04f34b-bb66-4247-92e4-c447085d9144\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.685006 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.18498592 +0000 UTC m=+153.541498129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.685574 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hsz8c\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.685630 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-plugins-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.685672 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53a6c85d-8623-46b6-9d53-860f1ff80f00-apiservice-cert\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.685715 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdnrc\" (UniqueName: \"kubernetes.io/projected/396991bf-e7c0-448e-a2fc-a626f0bd43b6-kube-api-access-fdnrc\") pod \"kube-storage-version-migrator-operator-b67b599dd-vgppb\" (UID: \"396991bf-e7c0-448e-a2fc-a626f0bd43b6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.702057 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3477fcf-270d-4b95-a581-3cfe8770cd0d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r4pgh\" (UID: \"f3477fcf-270d-4b95-a581-3cfe8770cd0d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.702143 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtcp2\" (UniqueName: \"kubernetes.io/projected/f421de15-2d9f-46ef-82da-414130bdf02e-kube-api-access-dtcp2\") pod \"multus-admission-controller-857f4d67dd-s2td2\" (UID: \"f421de15-2d9f-46ef-82da-414130bdf02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.702311 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ab99d3e-35d5-4a34-b195-fc53434b5168-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmjcj\" (UID: \"1ab99d3e-35d5-4a34-b195-fc53434b5168\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.702490 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42040525-2c1c-46bf-94e6-b23e05b7b3ae-installation-pull-secrets\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.702640 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/45bb8c11-1f6e-4f8a-a67d-b2f531211ba4-metrics-tls\") pod \"dns-default-hxhhw\" (UID: \"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4\") " pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.702647 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53a6c85d-8623-46b6-9d53-860f1ff80f00-apiservice-cert\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.702918 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/640d1c70-7a52-4f0d-a4df-a65d4b247acb-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hpqqz\" (UID: \"640d1c70-7a52-4f0d-a4df-a65d4b247acb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.702970 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c35b8c1e-cb44-4e0b-9732-0132cdb4e10a-serving-cert\") pod \"service-ca-operator-777779d784-qww9c\" (UID: \"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.702991 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws246\" (UniqueName: \"kubernetes.io/projected/c35b8c1e-cb44-4e0b-9732-0132cdb4e10a-kube-api-access-ws246\") pod \"service-ca-operator-777779d784-qww9c\" (UID: \"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703009 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdfxg\" (UniqueName: \"kubernetes.io/projected/c036f4c8-f544-4f0d-b1c6-51944420430c-kube-api-access-jdfxg\") pod \"collect-profiles-29401785-v5cgr\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703029 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-mountpoint-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703065 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5af8ff72-8043-4f10-8357-10ce2ff5123b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703094 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3477fcf-270d-4b95-a581-3cfe8770cd0d-config\") pod \"kube-controller-manager-operator-78b949d7b-r4pgh\" (UID: \"f3477fcf-270d-4b95-a581-3cfe8770cd0d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703118 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c3c979d-1a74-40a6-b24b-572c0ab97f63-config\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703136 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrwfw\" (UniqueName: \"kubernetes.io/projected/1da80646-0675-4f6c-ac13-3e19f44ce6a3-kube-api-access-qrwfw\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703153 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/842021b8-6f3c-4eab-bcba-0b27f6736e9e-images\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703174 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/53a6c85d-8623-46b6-9d53-860f1ff80f00-tmpfs\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703192 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9hg4\" (UniqueName: \"kubernetes.io/projected/c2a5d14a-4296-4064-8267-cf447ce6e41f-kube-api-access-x9hg4\") pod \"marketplace-operator-79b997595-hsz8c\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703237 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2b04f34b-bb66-4247-92e4-c447085d9144-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9wxs5\" (UID: \"2b04f34b-bb66-4247-92e4-c447085d9144\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703259 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6110aafd-c076-4fa4-a0b6-f5f434b9cfea-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wnjvs\" (UID: \"6110aafd-c076-4fa4-a0b6-f5f434b9cfea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703275 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8c3c979d-1a74-40a6-b24b-572c0ab97f63-etcd-client\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703291 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-csi-data-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703316 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hsz8c\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.703371 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-479lw\" (UniqueName: \"kubernetes.io/projected/b68a4880-028d-4379-ac82-67d9a9f6a1d3-kube-api-access-479lw\") pod \"migrator-59844c95c7-bc87m\" (UID: \"b68a4880-028d-4379-ac82-67d9a9f6a1d3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804016 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804175 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9fad68f4-2997-42c8-aeb2-937d220f7345-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kxjw7\" (UID: \"9fad68f4-2997-42c8-aeb2-937d220f7345\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804223 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c3c979d-1a74-40a6-b24b-572c0ab97f63-serving-cert\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804245 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c3c979d-1a74-40a6-b24b-572c0ab97f63-etcd-service-ca\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804268 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6110aafd-c076-4fa4-a0b6-f5f434b9cfea-config\") pod \"kube-apiserver-operator-766d6c64bb-wnjvs\" (UID: \"6110aafd-c076-4fa4-a0b6-f5f434b9cfea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804291 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45bb8c11-1f6e-4f8a-a67d-b2f531211ba4-config-volume\") pod \"dns-default-hxhhw\" (UID: \"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4\") " pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804348 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-socket-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804372 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zr9q\" (UniqueName: \"kubernetes.io/projected/842021b8-6f3c-4eab-bcba-0b27f6736e9e-kube-api-access-9zr9q\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804395 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9fad68f4-2997-42c8-aeb2-937d220f7345-proxy-tls\") pod \"machine-config-controller-84d6567774-kxjw7\" (UID: \"9fad68f4-2997-42c8-aeb2-937d220f7345\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804416 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-registration-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804446 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e46e33eb-d1b7-4f60-8911-23a2227776de-node-bootstrap-token\") pod \"machine-config-server-d7z6d\" (UID: \"e46e33eb-d1b7-4f60-8911-23a2227776de\") " pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804500 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/640d1c70-7a52-4f0d-a4df-a65d4b247acb-srv-cert\") pod \"olm-operator-6b444d44fb-hpqqz\" (UID: \"640d1c70-7a52-4f0d-a4df-a65d4b247acb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804542 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f421de15-2d9f-46ef-82da-414130bdf02e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s2td2\" (UID: \"f421de15-2d9f-46ef-82da-414130bdf02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804565 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/35c9fc97-1683-4083-9090-d1cfcc0a605e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bmxc6\" (UID: \"35c9fc97-1683-4083-9090-d1cfcc0a605e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804591 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/842021b8-6f3c-4eab-bcba-0b27f6736e9e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804640 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl5fz\" (UniqueName: \"kubernetes.io/projected/8c3c979d-1a74-40a6-b24b-572c0ab97f63-kube-api-access-pl5fz\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804664 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7ct7\" (UniqueName: \"kubernetes.io/projected/9fad68f4-2997-42c8-aeb2-937d220f7345-kube-api-access-b7ct7\") pod \"machine-config-controller-84d6567774-kxjw7\" (UID: \"9fad68f4-2997-42c8-aeb2-937d220f7345\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804717 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6110aafd-c076-4fa4-a0b6-f5f434b9cfea-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wnjvs\" (UID: \"6110aafd-c076-4fa4-a0b6-f5f434b9cfea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804744 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3477fcf-270d-4b95-a581-3cfe8770cd0d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r4pgh\" (UID: \"f3477fcf-270d-4b95-a581-3cfe8770cd0d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804801 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c036f4c8-f544-4f0d-b1c6-51944420430c-secret-volume\") pod \"collect-profiles-29401785-v5cgr\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804827 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5af8ff72-8043-4f10-8357-10ce2ff5123b-metrics-tls\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804869 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fz65\" (UniqueName: \"kubernetes.io/projected/fddf1692-d58b-4d3e-88c3-995e2c606b3b-kube-api-access-2fz65\") pod \"ingress-canary-pr9hq\" (UID: \"fddf1692-d58b-4d3e-88c3-995e2c606b3b\") " pod="openshift-ingress-canary/ingress-canary-pr9hq" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804958 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrtmj\" (UniqueName: \"kubernetes.io/projected/2b04f34b-bb66-4247-92e4-c447085d9144-kube-api-access-jrtmj\") pod \"openshift-config-operator-7777fb866f-9wxs5\" (UID: \"2b04f34b-bb66-4247-92e4-c447085d9144\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.804985 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnzgk\" (UniqueName: \"kubernetes.io/projected/35c9fc97-1683-4083-9090-d1cfcc0a605e-kube-api-access-hnzgk\") pod \"package-server-manager-789f6589d5-bmxc6\" (UID: \"35c9fc97-1683-4083-9090-d1cfcc0a605e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805010 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842021b8-6f3c-4eab-bcba-0b27f6736e9e-config\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805063 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8c3c979d-1a74-40a6-b24b-572c0ab97f63-etcd-ca\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805097 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b04f34b-bb66-4247-92e4-c447085d9144-serving-cert\") pod \"openshift-config-operator-7777fb866f-9wxs5\" (UID: \"2b04f34b-bb66-4247-92e4-c447085d9144\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805123 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hsz8c\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805147 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-plugins-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805208 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fddf1692-d58b-4d3e-88c3-995e2c606b3b-cert\") pod \"ingress-canary-pr9hq\" (UID: \"fddf1692-d58b-4d3e-88c3-995e2c606b3b\") " pod="openshift-ingress-canary/ingress-canary-pr9hq" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805250 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3477fcf-270d-4b95-a581-3cfe8770cd0d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r4pgh\" (UID: \"f3477fcf-270d-4b95-a581-3cfe8770cd0d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805275 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtcp2\" (UniqueName: \"kubernetes.io/projected/f421de15-2d9f-46ef-82da-414130bdf02e-kube-api-access-dtcp2\") pod \"multus-admission-controller-857f4d67dd-s2td2\" (UID: \"f421de15-2d9f-46ef-82da-414130bdf02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805324 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/45bb8c11-1f6e-4f8a-a67d-b2f531211ba4-metrics-tls\") pod \"dns-default-hxhhw\" (UID: \"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4\") " pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805350 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/640d1c70-7a52-4f0d-a4df-a65d4b247acb-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hpqqz\" (UID: \"640d1c70-7a52-4f0d-a4df-a65d4b247acb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805374 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e46e33eb-d1b7-4f60-8911-23a2227776de-certs\") pod \"machine-config-server-d7z6d\" (UID: \"e46e33eb-d1b7-4f60-8911-23a2227776de\") " pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805400 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb7vl\" (UniqueName: \"kubernetes.io/projected/e46e33eb-d1b7-4f60-8911-23a2227776de-kube-api-access-lb7vl\") pod \"machine-config-server-d7z6d\" (UID: \"e46e33eb-d1b7-4f60-8911-23a2227776de\") " pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805442 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdfxg\" (UniqueName: \"kubernetes.io/projected/c036f4c8-f544-4f0d-b1c6-51944420430c-kube-api-access-jdfxg\") pod \"collect-profiles-29401785-v5cgr\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805465 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-mountpoint-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805503 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5af8ff72-8043-4f10-8357-10ce2ff5123b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805529 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3477fcf-270d-4b95-a581-3cfe8770cd0d-config\") pod \"kube-controller-manager-operator-78b949d7b-r4pgh\" (UID: \"f3477fcf-270d-4b95-a581-3cfe8770cd0d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805562 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c3c979d-1a74-40a6-b24b-572c0ab97f63-config\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805585 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrwfw\" (UniqueName: \"kubernetes.io/projected/1da80646-0675-4f6c-ac13-3e19f44ce6a3-kube-api-access-qrwfw\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805607 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/842021b8-6f3c-4eab-bcba-0b27f6736e9e-images\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805633 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9hg4\" (UniqueName: \"kubernetes.io/projected/c2a5d14a-4296-4064-8267-cf447ce6e41f-kube-api-access-x9hg4\") pod \"marketplace-operator-79b997595-hsz8c\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805657 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2b04f34b-bb66-4247-92e4-c447085d9144-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9wxs5\" (UID: \"2b04f34b-bb66-4247-92e4-c447085d9144\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805677 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6110aafd-c076-4fa4-a0b6-f5f434b9cfea-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wnjvs\" (UID: \"6110aafd-c076-4fa4-a0b6-f5f434b9cfea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805700 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8c3c979d-1a74-40a6-b24b-572c0ab97f63-etcd-client\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805719 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-csi-data-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805741 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hsz8c\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805763 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-479lw\" (UniqueName: \"kubernetes.io/projected/b68a4880-028d-4379-ac82-67d9a9f6a1d3-kube-api-access-479lw\") pod \"migrator-59844c95c7-bc87m\" (UID: \"b68a4880-028d-4379-ac82-67d9a9f6a1d3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805812 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpc99\" (UniqueName: \"kubernetes.io/projected/5af8ff72-8043-4f10-8357-10ce2ff5123b-kube-api-access-bpc99\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805869 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c036f4c8-f544-4f0d-b1c6-51944420430c-config-volume\") pod \"collect-profiles-29401785-v5cgr\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805914 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8sv8\" (UniqueName: \"kubernetes.io/projected/45bb8c11-1f6e-4f8a-a67d-b2f531211ba4-kube-api-access-w8sv8\") pod \"dns-default-hxhhw\" (UID: \"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4\") " pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805954 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-624ws\" (UniqueName: \"kubernetes.io/projected/640d1c70-7a52-4f0d-a4df-a65d4b247acb-kube-api-access-624ws\") pod \"olm-operator-6b444d44fb-hpqqz\" (UID: \"640d1c70-7a52-4f0d-a4df-a65d4b247acb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.805978 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5af8ff72-8043-4f10-8357-10ce2ff5123b-trusted-ca\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.809140 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.309084623 +0000 UTC m=+153.665596782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.809614 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-socket-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.809861 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-registration-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.809897 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-plugins-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.810052 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-csi-data-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.810238 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1da80646-0675-4f6c-ac13-3e19f44ce6a3-mountpoint-dir\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.881791 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2b04f34b-bb66-4247-92e4-c447085d9144-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9wxs5\" (UID: \"2b04f34b-bb66-4247-92e4-c447085d9144\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.882118 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/53a6c85d-8623-46b6-9d53-860f1ff80f00-tmpfs\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.883493 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42040525-2c1c-46bf-94e6-b23e05b7b3ae-ca-trust-extracted\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.883610 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5af8ff72-8043-4f10-8357-10ce2ff5123b-trusted-ca\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.884446 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842021b8-6f3c-4eab-bcba-0b27f6736e9e-config\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.886771 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/842021b8-6f3c-4eab-bcba-0b27f6736e9e-images\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.891177 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c35b8c1e-cb44-4e0b-9732-0132cdb4e10a-config\") pod \"service-ca-operator-777779d784-qww9c\" (UID: \"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.897553 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9fad68f4-2997-42c8-aeb2-937d220f7345-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kxjw7\" (UID: \"9fad68f4-2997-42c8-aeb2-937d220f7345\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.900740 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-trusted-ca\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.901847 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c3c979d-1a74-40a6-b24b-572c0ab97f63-serving-cert\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.902015 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6110aafd-c076-4fa4-a0b6-f5f434b9cfea-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wnjvs\" (UID: \"6110aafd-c076-4fa4-a0b6-f5f434b9cfea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.902673 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-certificates\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.903182 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9fad68f4-2997-42c8-aeb2-937d220f7345-proxy-tls\") pod \"machine-config-controller-84d6567774-kxjw7\" (UID: \"9fad68f4-2997-42c8-aeb2-937d220f7345\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.904526 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f421de15-2d9f-46ef-82da-414130bdf02e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s2td2\" (UID: \"f421de15-2d9f-46ef-82da-414130bdf02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.905255 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c036f4c8-f544-4f0d-b1c6-51944420430c-config-volume\") pod \"collect-profiles-29401785-v5cgr\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.890939 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c3c979d-1a74-40a6-b24b-572c0ab97f63-config\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.906121 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8c3c979d-1a74-40a6-b24b-572c0ab97f63-etcd-ca\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.906789 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t65tp\" (UniqueName: \"kubernetes.io/projected/53a6c85d-8623-46b6-9d53-860f1ff80f00-kube-api-access-t65tp\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.906844 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hsz8c\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.908194 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e46e33eb-d1b7-4f60-8911-23a2227776de-node-bootstrap-token\") pod \"machine-config-server-d7z6d\" (UID: \"e46e33eb-d1b7-4f60-8911-23a2227776de\") " pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.908269 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.908772 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ab99d3e-35d5-4a34-b195-fc53434b5168-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmjcj\" (UID: \"1ab99d3e-35d5-4a34-b195-fc53434b5168\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.909347 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5af8ff72-8043-4f10-8357-10ce2ff5123b-metrics-tls\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.909430 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fz65\" (UniqueName: \"kubernetes.io/projected/fddf1692-d58b-4d3e-88c3-995e2c606b3b-kube-api-access-2fz65\") pod \"ingress-canary-pr9hq\" (UID: \"fddf1692-d58b-4d3e-88c3-995e2c606b3b\") " pod="openshift-ingress-canary/ingress-canary-pr9hq" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.910439 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fddf1692-d58b-4d3e-88c3-995e2c606b3b-cert\") pod \"ingress-canary-pr9hq\" (UID: \"fddf1692-d58b-4d3e-88c3-995e2c606b3b\") " pod="openshift-ingress-canary/ingress-canary-pr9hq" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.910625 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e46e33eb-d1b7-4f60-8911-23a2227776de-certs\") pod \"machine-config-server-d7z6d\" (UID: \"e46e33eb-d1b7-4f60-8911-23a2227776de\") " pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.910679 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb7vl\" (UniqueName: \"kubernetes.io/projected/e46e33eb-d1b7-4f60-8911-23a2227776de-kube-api-access-lb7vl\") pod \"machine-config-server-d7z6d\" (UID: \"e46e33eb-d1b7-4f60-8911-23a2227776de\") " pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:34 crc kubenswrapper[4931]: E1125 21:53:34.910766 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.41074327 +0000 UTC m=+153.767255419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.917978 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/842021b8-6f3c-4eab-bcba-0b27f6736e9e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.919342 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/396991bf-e7c0-448e-a2fc-a626f0bd43b6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vgppb\" (UID: \"396991bf-e7c0-448e-a2fc-a626f0bd43b6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.919850 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/35c9fc97-1683-4083-9090-d1cfcc0a605e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bmxc6\" (UID: \"35c9fc97-1683-4083-9090-d1cfcc0a605e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.920689 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/640d1c70-7a52-4f0d-a4df-a65d4b247acb-srv-cert\") pod \"olm-operator-6b444d44fb-hpqqz\" (UID: \"640d1c70-7a52-4f0d-a4df-a65d4b247acb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.920821 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53a6c85d-8623-46b6-9d53-860f1ff80f00-webhook-cert\") pod \"packageserver-d55dfcdfc-2gvlz\" (UID: \"53a6c85d-8623-46b6-9d53-860f1ff80f00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.921633 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3477fcf-270d-4b95-a581-3cfe8770cd0d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r4pgh\" (UID: \"f3477fcf-270d-4b95-a581-3cfe8770cd0d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.922175 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8c3c979d-1a74-40a6-b24b-572c0ab97f63-etcd-client\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.922654 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b04f34b-bb66-4247-92e4-c447085d9144-serving-cert\") pod \"openshift-config-operator-7777fb866f-9wxs5\" (UID: \"2b04f34b-bb66-4247-92e4-c447085d9144\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.927450 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42040525-2c1c-46bf-94e6-b23e05b7b3ae-installation-pull-secrets\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.927866 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fddf1692-d58b-4d3e-88c3-995e2c606b3b-cert\") pod \"ingress-canary-pr9hq\" (UID: \"fddf1692-d58b-4d3e-88c3-995e2c606b3b\") " pod="openshift-ingress-canary/ingress-canary-pr9hq" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.928163 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/640d1c70-7a52-4f0d-a4df-a65d4b247acb-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hpqqz\" (UID: \"640d1c70-7a52-4f0d-a4df-a65d4b247acb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.928672 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7ct7\" (UniqueName: \"kubernetes.io/projected/9fad68f4-2997-42c8-aeb2-937d220f7345-kube-api-access-b7ct7\") pod \"machine-config-controller-84d6567774-kxjw7\" (UID: \"9fad68f4-2997-42c8-aeb2-937d220f7345\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.929688 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bf866554-0f21-4262-9b8e-32d9c0964488-profile-collector-cert\") pod \"catalog-operator-68c6474976-gkwhd\" (UID: \"bf866554-0f21-4262-9b8e-32d9c0964488\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.930469 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl5fz\" (UniqueName: \"kubernetes.io/projected/8c3c979d-1a74-40a6-b24b-572c0ab97f63-kube-api-access-pl5fz\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.930737 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-bound-sa-token\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.932372 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45bb8c11-1f6e-4f8a-a67d-b2f531211ba4-config-volume\") pod \"dns-default-hxhhw\" (UID: \"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4\") " pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.932777 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-tls\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.935409 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf6f8\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-kube-api-access-pf6f8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.936667 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6110aafd-c076-4fa4-a0b6-f5f434b9cfea-config\") pod \"kube-apiserver-operator-766d6c64bb-wnjvs\" (UID: \"6110aafd-c076-4fa4-a0b6-f5f434b9cfea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.936749 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hsz8c\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.937428 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws246\" (UniqueName: \"kubernetes.io/projected/c35b8c1e-cb44-4e0b-9732-0132cdb4e10a-kube-api-access-ws246\") pod \"service-ca-operator-777779d784-qww9c\" (UID: \"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.937706 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c35b8c1e-cb44-4e0b-9732-0132cdb4e10a-serving-cert\") pod \"service-ca-operator-777779d784-qww9c\" (UID: \"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.938571 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3477fcf-270d-4b95-a581-3cfe8770cd0d-config\") pod \"kube-controller-manager-operator-78b949d7b-r4pgh\" (UID: \"f3477fcf-270d-4b95-a581-3cfe8770cd0d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.939314 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e46e33eb-d1b7-4f60-8911-23a2227776de-certs\") pod \"machine-config-server-d7z6d\" (UID: \"e46e33eb-d1b7-4f60-8911-23a2227776de\") " pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.940038 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6110aafd-c076-4fa4-a0b6-f5f434b9cfea-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wnjvs\" (UID: \"6110aafd-c076-4fa4-a0b6-f5f434b9cfea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.941105 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c036f4c8-f544-4f0d-b1c6-51944420430c-secret-volume\") pod \"collect-profiles-29401785-v5cgr\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.943522 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/45bb8c11-1f6e-4f8a-a67d-b2f531211ba4-metrics-tls\") pod \"dns-default-hxhhw\" (UID: \"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4\") " pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.943861 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/396991bf-e7c0-448e-a2fc-a626f0bd43b6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vgppb\" (UID: \"396991bf-e7c0-448e-a2fc-a626f0bd43b6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.950934 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e46e33eb-d1b7-4f60-8911-23a2227776de-node-bootstrap-token\") pod \"machine-config-server-d7z6d\" (UID: \"e46e33eb-d1b7-4f60-8911-23a2227776de\") " pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.955794 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdnrc\" (UniqueName: \"kubernetes.io/projected/396991bf-e7c0-448e-a2fc-a626f0bd43b6-kube-api-access-fdnrc\") pod \"kube-storage-version-migrator-operator-b67b599dd-vgppb\" (UID: \"396991bf-e7c0-448e-a2fc-a626f0bd43b6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.959270 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrtmj\" (UniqueName: \"kubernetes.io/projected/2b04f34b-bb66-4247-92e4-c447085d9144-kube-api-access-jrtmj\") pod \"openshift-config-operator-7777fb866f-9wxs5\" (UID: \"2b04f34b-bb66-4247-92e4-c447085d9144\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.978820 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bf866554-0f21-4262-9b8e-32d9c0964488-srv-cert\") pod \"catalog-operator-68c6474976-gkwhd\" (UID: \"bf866554-0f21-4262-9b8e-32d9c0964488\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.979285 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wfjj\" (UniqueName: \"kubernetes.io/projected/bf866554-0f21-4262-9b8e-32d9c0964488-kube-api-access-7wfjj\") pod \"catalog-operator-68c6474976-gkwhd\" (UID: \"bf866554-0f21-4262-9b8e-32d9c0964488\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.983951 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3477fcf-270d-4b95-a581-3cfe8770cd0d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r4pgh\" (UID: \"f3477fcf-270d-4b95-a581-3cfe8770cd0d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.985253 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c3c979d-1a74-40a6-b24b-572c0ab97f63-etcd-service-ca\") pod \"etcd-operator-b45778765-6vpzc\" (UID: \"8c3c979d-1a74-40a6-b24b-572c0ab97f63\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.985322 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ab99d3e-35d5-4a34-b195-fc53434b5168-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmjcj\" (UID: \"1ab99d3e-35d5-4a34-b195-fc53434b5168\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.986421 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ab99d3e-35d5-4a34-b195-fc53434b5168-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-gmjcj\" (UID: \"1ab99d3e-35d5-4a34-b195-fc53434b5168\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.994543 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:34 crc kubenswrapper[4931]: I1125 21:53:34.998833 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-479lw\" (UniqueName: \"kubernetes.io/projected/b68a4880-028d-4379-ac82-67d9a9f6a1d3-kube-api-access-479lw\") pod \"migrator-59844c95c7-bc87m\" (UID: \"b68a4880-028d-4379-ac82-67d9a9f6a1d3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.012312 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.012924 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.512906059 +0000 UTC m=+153.869418188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.016217 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrwfw\" (UniqueName: \"kubernetes.io/projected/1da80646-0675-4f6c-ac13-3e19f44ce6a3-kube-api-access-qrwfw\") pod \"csi-hostpathplugin-2c4n7\" (UID: \"1da80646-0675-4f6c-ac13-3e19f44ce6a3\") " pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.029791 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.049472 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnzgk\" (UniqueName: \"kubernetes.io/projected/35c9fc97-1683-4083-9090-d1cfcc0a605e-kube-api-access-hnzgk\") pod \"package-server-manager-789f6589d5-bmxc6\" (UID: \"35c9fc97-1683-4083-9090-d1cfcc0a605e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.064494 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9hg4\" (UniqueName: \"kubernetes.io/projected/c2a5d14a-4296-4064-8267-cf447ce6e41f-kube-api-access-x9hg4\") pod \"marketplace-operator-79b997595-hsz8c\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.078156 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zr9q\" (UniqueName: \"kubernetes.io/projected/842021b8-6f3c-4eab-bcba-0b27f6736e9e-kube-api-access-9zr9q\") pod \"machine-api-operator-5694c8668f-5dlk4\" (UID: \"842021b8-6f3c-4eab-bcba-0b27f6736e9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.078262 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.084932 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.095127 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.098989 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8sv8\" (UniqueName: \"kubernetes.io/projected/45bb8c11-1f6e-4f8a-a67d-b2f531211ba4-kube-api-access-w8sv8\") pod \"dns-default-hxhhw\" (UID: \"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4\") " pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.099262 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.105853 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.114222 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.114583 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.614567614 +0000 UTC m=+153.971079733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.116978 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpc99\" (UniqueName: \"kubernetes.io/projected/5af8ff72-8043-4f10-8357-10ce2ff5123b-kube-api-access-bpc99\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.127023 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.132219 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-624ws\" (UniqueName: \"kubernetes.io/projected/640d1c70-7a52-4f0d-a4df-a65d4b247acb-kube-api-access-624ws\") pod \"olm-operator-6b444d44fb-hpqqz\" (UID: \"640d1c70-7a52-4f0d-a4df-a65d4b247acb\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.150621 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.158631 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5af8ff72-8043-4f10-8357-10ce2ff5123b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tp4r9\" (UID: \"5af8ff72-8043-4f10-8357-10ce2ff5123b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.166749 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.176940 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdfxg\" (UniqueName: \"kubernetes.io/projected/c036f4c8-f544-4f0d-b1c6-51944420430c-kube-api-access-jdfxg\") pod \"collect-profiles-29401785-v5cgr\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.180255 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.188475 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.191982 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtcp2\" (UniqueName: \"kubernetes.io/projected/f421de15-2d9f-46ef-82da-414130bdf02e-kube-api-access-dtcp2\") pod \"multus-admission-controller-857f4d67dd-s2td2\" (UID: \"f421de15-2d9f-46ef-82da-414130bdf02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.216211 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.216385 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.716361714 +0000 UTC m=+154.072873833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.216482 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.216980 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.716974001 +0000 UTC m=+154.073486120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.219110 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.227105 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.262815 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb7vl\" (UniqueName: \"kubernetes.io/projected/e46e33eb-d1b7-4f60-8911-23a2227776de-kube-api-access-lb7vl\") pod \"machine-config-server-d7z6d\" (UID: \"e46e33eb-d1b7-4f60-8911-23a2227776de\") " pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.268688 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st"] Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.276714 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fz65\" (UniqueName: \"kubernetes.io/projected/fddf1692-d58b-4d3e-88c3-995e2c606b3b-kube-api-access-2fz65\") pod \"ingress-canary-pr9hq\" (UID: \"fddf1692-d58b-4d3e-88c3-995e2c606b3b\") " pod="openshift-ingress-canary/ingress-canary-pr9hq" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.283644 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.304436 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5"] Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.310021 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-d7z6d" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.315485 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.317563 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.318162 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.318226 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.318263 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.318291 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.818262337 +0000 UTC m=+154.174774456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.318316 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.318869 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.321855 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.326305 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0211b0cb-4e13-45ce-aab7-a0e6848c2bef-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rwnnf\" (UID: \"0211b0cb-4e13-45ce-aab7-a0e6848c2bef\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.329040 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.330380 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-p4p2g\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.382088 4931 generic.go:334] "Generic (PLEG): container finished" podID="9d1e5baf-648e-4327-afc5-7a2835d7fb21" containerID="9d5021f1f74b366cd3465e4a29667cf03d654285958153e7c03a56161492519f" exitCode=0 Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.382195 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" event={"ID":"9d1e5baf-648e-4327-afc5-7a2835d7fb21","Type":"ContainerDied","Data":"9d5021f1f74b366cd3465e4a29667cf03d654285958153e7c03a56161492519f"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.389357 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" event={"ID":"fada681c-6ade-49e4-b324-1985853bc61e","Type":"ContainerStarted","Data":"e07c9a9d080ea1c3fd5e19dd6f5e4bebdc621177a01e5c9fd3d9d377bac21c13"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.392697 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2hpxq" event={"ID":"69dfd797-b979-42a6-a4d5-d95e48bf7aa2","Type":"ContainerStarted","Data":"2375b11e4468b35b3222b4c91592af8c3a6668a8439cb2fa332fd8a374702738"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.394064 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vqmcg" event={"ID":"b4dd52d7-4597-4159-a728-012cd2180b35","Type":"ContainerStarted","Data":"d2a061c956ab5b7bc0910d7438c19a4709f08ca4442bdef7a75fc8a8cf77c5a3"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.395276 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" event={"ID":"74cbb6cf-4717-439c-b712-67718c154742","Type":"ContainerStarted","Data":"65181b64fc25f693bd5d81c886a8293a876f472850faf7995121ecaa6c7b1942"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.412253 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.414870 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" event={"ID":"07b1ae10-2353-4d92-850b-2370581c674f","Type":"ContainerStarted","Data":"f5d976fe09269413270d30a8b8f5d076dfd9941d49d1495c72fe180babdde197"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.417322 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" event={"ID":"b1c07b60-6159-4cbd-a62c-d684bd8dbd22","Type":"ContainerStarted","Data":"072597adaa4ede933cc5266a901eafc2697bedd6b83e0373bc4524ba8161127a"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.419546 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.420209 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:35.920190669 +0000 UTC m=+154.276702788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.420195 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" event={"ID":"ffaf251e-ab8f-4339-905d-5bd8f700beee","Type":"ContainerStarted","Data":"fdef55b511b3fc1948b077bde93dd00de70e964a21ddb3ff1535340fbfd435c9"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.421748 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" event={"ID":"45811a4b-98c3-4f54-b772-f1adc68d88c7","Type":"ContainerStarted","Data":"7c5a9184b82ab596290f82d0f41004294c4ff29ce3c3989db14b27a8fe4fc73c"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.424747 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" event={"ID":"55263751-50ce-4c6b-a3e9-25c7fee34e37","Type":"ContainerStarted","Data":"798ba29c10e7825cb4ad618fcb49208f6b848db808dc921dfdbe97a3428f9577"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.427942 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gdm7n" event={"ID":"ec936935-d216-4775-be35-f45ed5c5ceca","Type":"ContainerStarted","Data":"314d8fe38b81acb19066940bfd883ab5ee3693b46c421bf569946fa866798c34"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.429779 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" event={"ID":"681adfb6-f6a8-4934-ade0-a1499bf95d2f","Type":"ContainerStarted","Data":"79eeef339e572963ff0eb3e131ddf8c37888cbe4dc7ee9be2b02542c29271dbf"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.431117 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9dxf2" event={"ID":"7445c312-10f0-4296-ba35-2320e22382e8","Type":"ContainerStarted","Data":"8576fc2320aa7a788f2ffa24892f7f7ab4b1cf3fb07298f88e73a6a896b4f7e3"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.432071 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9dxf2" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.433659 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.433715 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.441108 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" event={"ID":"32a6e5f5-b72c-4e07-a87c-72d7420a1599","Type":"ContainerStarted","Data":"9d633d3e976b2814e3cddaa69f4d9855724c1f7372ab50a53662a2e4f6de51db"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.443540 4931 generic.go:334] "Generic (PLEG): container finished" podID="3151332c-78d1-439b-bf22-25e60ed60dd9" containerID="dcd5da91a7e97bc4e25040d162353a86b27d24016396f48b3b40a90de3df2ef6" exitCode=0 Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.444153 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" event={"ID":"3151332c-78d1-439b-bf22-25e60ed60dd9","Type":"ContainerDied","Data":"dcd5da91a7e97bc4e25040d162353a86b27d24016396f48b3b40a90de3df2ef6"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.444925 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m"] Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.446442 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" event={"ID":"54f4b8f7-8d9c-4bef-af7f-cb2335639e50","Type":"ContainerStarted","Data":"6ea6beea3ae6cf523fa5d81713c30abe206daba8eed87b40e8d51e61be8c2149"} Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.451848 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.460701 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.473625 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.517919 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.521254 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.523015 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.022826843 +0000 UTC m=+154.379338972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.536448 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pr9hq" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.630070 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.632549 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.132531319 +0000 UTC m=+154.489043518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.663487 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs"] Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.710016 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz"] Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.733267 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.733436 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.233411384 +0000 UTC m=+154.589923513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.733737 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.734128 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.234119534 +0000 UTC m=+154.590631713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.836618 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.836972 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.336946281 +0000 UTC m=+154.693458400 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.837131 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.838175 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.338124744 +0000 UTC m=+154.694636863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.938158 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.938310 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.438284338 +0000 UTC m=+154.794796467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.938716 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:35 crc kubenswrapper[4931]: E1125 21:53:35.939068 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.439053749 +0000 UTC m=+154.795565868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.960384 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.962681 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 25 21:53:35 crc kubenswrapper[4931]: I1125 21:53:35.962727 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.005044 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.007962 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.039775 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.040071 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.540057497 +0000 UTC m=+154.896569616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.072838 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.073455 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2c4n7"] Nov 25 21:53:36 crc kubenswrapper[4931]: W1125 21:53:36.138100 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ab99d3e_35d5_4a34_b195_fc53434b5168.slice/crio-68e08f99a93aac7eac5bc35e6a03901593f7e1b8767fb2d36e0bb364f8cccc17 WatchSource:0}: Error finding container 68e08f99a93aac7eac5bc35e6a03901593f7e1b8767fb2d36e0bb364f8cccc17: Status 404 returned error can't find the container with id 68e08f99a93aac7eac5bc35e6a03901593f7e1b8767fb2d36e0bb364f8cccc17 Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.144619 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.145275 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.64525777 +0000 UTC m=+155.001769889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: W1125 21:53:36.174780 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod396991bf_e7c0_448e_a2fc_a626f0bd43b6.slice/crio-94b3864fba33f12a430d0efc9fd45951cbf063fb410a43736353b713d9966a80 WatchSource:0}: Error finding container 94b3864fba33f12a430d0efc9fd45951cbf063fb410a43736353b713d9966a80: Status 404 returned error can't find the container with id 94b3864fba33f12a430d0efc9fd45951cbf063fb410a43736353b713d9966a80 Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.228121 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.240457 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.245395 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.245663 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.745632261 +0000 UTC m=+155.102144380 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.247629 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qww9c"] Nov 25 21:53:36 crc kubenswrapper[4931]: W1125 21:53:36.322475 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod640d1c70_7a52_4f0d_a4df_a65d4b247acb.slice/crio-73b086613889aa193b7e346d6629584e5b44a35374a3935ed28d4e804d7f6506 WatchSource:0}: Error finding container 73b086613889aa193b7e346d6629584e5b44a35374a3935ed28d4e804d7f6506: Status 404 returned error can't find the container with id 73b086613889aa193b7e346d6629584e5b44a35374a3935ed28d4e804d7f6506 Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.347449 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.347973 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.847941694 +0000 UTC m=+155.204453813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.451615 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.453323 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.953299681 +0000 UTC m=+155.309811800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.453389 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.453858 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:36.953850337 +0000 UTC m=+155.310362456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.465668 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6vpzc"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.468474 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hsz8c"] Nov 25 21:53:36 crc kubenswrapper[4931]: W1125 21:53:36.482950 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c3c979d_1a74_40a6_b24b_572c0ab97f63.slice/crio-486950313a37c860c8ee1a0b7a44e1e8bdebc3bbf58ca6f258eeb25d1789c38f WatchSource:0}: Error finding container 486950313a37c860c8ee1a0b7a44e1e8bdebc3bbf58ca6f258eeb25d1789c38f: Status 404 returned error can't find the container with id 486950313a37c860c8ee1a0b7a44e1e8bdebc3bbf58ca6f258eeb25d1789c38f Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.496008 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" event={"ID":"1da80646-0675-4f6c-ac13-3e19f44ce6a3","Type":"ContainerStarted","Data":"e0a36dcc5ae5b44dabe0403be0e3eb779a29d947cc786e7f3f272fed4b98a638"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.498055 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" event={"ID":"1ab99d3e-35d5-4a34-b195-fc53434b5168","Type":"ContainerStarted","Data":"68e08f99a93aac7eac5bc35e6a03901593f7e1b8767fb2d36e0bb364f8cccc17"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.499217 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" event={"ID":"53a6c85d-8623-46b6-9d53-860f1ff80f00","Type":"ContainerStarted","Data":"5fbb16967afc36bc45fd7727a49493f9dd134ce2814e85705ff2834f689e5573"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.499801 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.500608 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" event={"ID":"45811a4b-98c3-4f54-b772-f1adc68d88c7","Type":"ContainerStarted","Data":"e27215a55f0ced530a2c15d0ef42f313d6649d6fb0de37272dd334eeea317714"} Nov 25 21:53:36 crc kubenswrapper[4931]: W1125 21:53:36.503969 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2a5d14a_4296_4064_8267_cf447ce6e41f.slice/crio-8d96b9e65d869ba2abccd66f54d696c92c730a7d4b08a98fda8cc2976eacaf3a WatchSource:0}: Error finding container 8d96b9e65d869ba2abccd66f54d696c92c730a7d4b08a98fda8cc2976eacaf3a: Status 404 returned error can't find the container with id 8d96b9e65d869ba2abccd66f54d696c92c730a7d4b08a98fda8cc2976eacaf3a Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.527666 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" event={"ID":"2b04f34b-bb66-4247-92e4-c447085d9144","Type":"ContainerStarted","Data":"7d4d16adbef6833081ef05098d7f45e5f1b817d8f1af47ff4b322f480b491429"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.527711 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" event={"ID":"2b04f34b-bb66-4247-92e4-c447085d9144","Type":"ContainerStarted","Data":"7214206ce5688eaae1c345ab0b1d1ffd72b59c20a1ef2ecc021609e23c8e8ccc"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.531567 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s2td2"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.531620 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vqmcg" event={"ID":"b4dd52d7-4597-4159-a728-012cd2180b35","Type":"ContainerStarted","Data":"6e5c7d633203e7a576cb3157de78969f44dd0134cbd73849288495745716decc"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.554597 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" event={"ID":"b1c07b60-6159-4cbd-a62c-d684bd8dbd22","Type":"ContainerStarted","Data":"df7571f090849216d162997506d1df7587b30c0b6fafe280906c048020acfd6e"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.554918 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.555285 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.055267706 +0000 UTC m=+155.411779825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.560187 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.560806 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.060786718 +0000 UTC m=+155.417298837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.567895 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.569508 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:36 crc kubenswrapper[4931]: W1125 21:53:36.569907 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3477fcf_270d_4b95_a581_3cfe8770cd0d.slice/crio-d760ebea0df9a1f2a8a8f15cb557cdd1d843ac3481edb6ce063ef304f73edacb WatchSource:0}: Error finding container d760ebea0df9a1f2a8a8f15cb557cdd1d843ac3481edb6ce063ef304f73edacb: Status 404 returned error can't find the container with id d760ebea0df9a1f2a8a8f15cb557cdd1d843ac3481edb6ce063ef304f73edacb Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.583321 4931 patch_prober.go:28] interesting pod/console-operator-58897d9998-vqmcg container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.583379 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-vqmcg" podUID="b4dd52d7-4597-4159-a728-012cd2180b35" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.589021 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.593705 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hxhhw"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.597918 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" event={"ID":"ffaf251e-ab8f-4339-905d-5bd8f700beee","Type":"ContainerStarted","Data":"bb8a2a6ec73095da179517739c5bd38fa9abcff1abc187af69a554a7515f2e2c"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.599187 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.601172 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" event={"ID":"681adfb6-f6a8-4934-ade0-a1499bf95d2f","Type":"ContainerStarted","Data":"f5ae9df9a292d15f6f9bd04337d1c1bd02dcbbf5b92ecb4a182b1d1558d0922a"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.601951 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pr9hq"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.603511 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" event={"ID":"bf866554-0f21-4262-9b8e-32d9c0964488","Type":"ContainerStarted","Data":"9a77803dab64df1dda43e4fcb5e2820835e106877b1175ed373d1ffcaad94558"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.604496 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.612706 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" event={"ID":"8385c8d3-791b-48c6-9173-84b036a6f294","Type":"ContainerStarted","Data":"dcb52d00bafb4561c6c460cad83fba94c752e054867e538693ee3362868da7c5"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.621552 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rsbdr" podStartSLOduration=132.621523228 podStartE2EDuration="2m12.621523228s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:36.615281446 +0000 UTC m=+154.971793566" watchObservedRunningTime="2025-11-25 21:53:36.621523228 +0000 UTC m=+154.978035347" Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.629559 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-d7z6d" event={"ID":"e46e33eb-d1b7-4f60-8911-23a2227776de","Type":"ContainerStarted","Data":"b3fc7b9e8d0162cefc9cc24b49d000bf654e04120f7a8fc2dcdae41e1b7985b3"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.649378 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m" event={"ID":"b68a4880-028d-4379-ac82-67d9a9f6a1d3","Type":"ContainerStarted","Data":"c803cd0ee40725f27a7bdc113482819fbe09f4f20d04e95d486bce1a61a8bc44"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.649465 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m" event={"ID":"b68a4880-028d-4379-ac82-67d9a9f6a1d3","Type":"ContainerStarted","Data":"43d1ec2babad0b13918a53237ca34fbe0c5a977e2e69c59192743e7d97215a2c"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.653952 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" event={"ID":"640d1c70-7a52-4f0d-a4df-a65d4b247acb","Type":"ContainerStarted","Data":"73b086613889aa193b7e346d6629584e5b44a35374a3935ed28d4e804d7f6506"} Nov 25 21:53:36 crc kubenswrapper[4931]: W1125 21:53:36.655381 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfddf1692_d58b_4d3e_88c3_995e2c606b3b.slice/crio-346ba04c441250a871894a3426b720bcd4b538fb36cae44dcf0041b77889d69b WatchSource:0}: Error finding container 346ba04c441250a871894a3426b720bcd4b538fb36cae44dcf0041b77889d69b: Status 404 returned error can't find the container with id 346ba04c441250a871894a3426b720bcd4b538fb36cae44dcf0041b77889d69b Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.657833 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" event={"ID":"fada681c-6ade-49e4-b324-1985853bc61e","Type":"ContainerStarted","Data":"7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.658657 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.662280 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" event={"ID":"54f4b8f7-8d9c-4bef-af7f-cb2335639e50","Type":"ContainerStarted","Data":"0b7617427094c8394c93048c7e10418917b7cf812c7afc86b13ac9098aeb49fe"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.663149 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.664246 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.164223612 +0000 UTC m=+155.520735781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.667806 4931 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-454st container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.667941 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" podUID="fada681c-6ade-49e4-b324-1985853bc61e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.676783 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" event={"ID":"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a","Type":"ContainerStarted","Data":"9fd7d4431f0bd372a6c77d2d57aaa024f9bfe70bbfac3b7d5a950e86611b243b"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.678740 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" event={"ID":"6110aafd-c076-4fa4-a0b6-f5f434b9cfea","Type":"ContainerStarted","Data":"f5270cb4700d346b9f8541e8d9395d6d942410f5c00b17e80a4691cdf21b6fec"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.680654 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" event={"ID":"9fad68f4-2997-42c8-aeb2-937d220f7345","Type":"ContainerStarted","Data":"dc6ec92efbbe837953ba76154698f276c186dad8b5f6853426dcdc3610ab3e53"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.682141 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" event={"ID":"396991bf-e7c0-448e-a2fc-a626f0bd43b6","Type":"ContainerStarted","Data":"94b3864fba33f12a430d0efc9fd45951cbf063fb410a43736353b713d9966a80"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.684939 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" event={"ID":"32a6e5f5-b72c-4e07-a87c-72d7420a1599","Type":"ContainerStarted","Data":"c17867efcefec5ca6b499e85f2e5441937696b9039cd6f51384d34b97188a53c"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.684997 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" event={"ID":"32a6e5f5-b72c-4e07-a87c-72d7420a1599","Type":"ContainerStarted","Data":"90fe57d36545ff8aa60e14fc1b056dc252846163a8deef1a21aceab88cb042ec"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.691219 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" event={"ID":"74cbb6cf-4717-439c-b712-67718c154742","Type":"ContainerStarted","Data":"98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd"} Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.692256 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.692325 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.759133 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5dlk4"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.764341 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.765783 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.265768775 +0000 UTC m=+155.622280884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.802023 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p4p2g"] Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.812149 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kxgml" podStartSLOduration=133.812133961 podStartE2EDuration="2m13.812133961s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:36.811005979 +0000 UTC m=+155.167518118" watchObservedRunningTime="2025-11-25 21:53:36.812133961 +0000 UTC m=+155.168646070" Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.866410 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.866689 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.3666715 +0000 UTC m=+155.723183619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:36 crc kubenswrapper[4931]: W1125 21:53:36.894534 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod842021b8_6f3c_4eab_bcba_0b27f6736e9e.slice/crio-d3047cb1968e9c75bc8f474d32ea059522b5d5a331f1d307a1bd55fc6aa1ac3f WatchSource:0}: Error finding container d3047cb1968e9c75bc8f474d32ea059522b5d5a331f1d307a1bd55fc6aa1ac3f: Status 404 returned error can't find the container with id d3047cb1968e9c75bc8f474d32ea059522b5d5a331f1d307a1bd55fc6aa1ac3f Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.935967 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-2hpxq" podStartSLOduration=132.935945485 podStartE2EDuration="2m12.935945485s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:36.899634527 +0000 UTC m=+155.256146646" watchObservedRunningTime="2025-11-25 21:53:36.935945485 +0000 UTC m=+155.292457604" Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.962061 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.962135 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 25 21:53:36 crc kubenswrapper[4931]: I1125 21:53:36.972306 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:36 crc kubenswrapper[4931]: E1125 21:53:36.972632 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.472616104 +0000 UTC m=+155.829128223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.073890 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.074034 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.574017423 +0000 UTC m=+155.930529542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.130765 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-wvj9s" podStartSLOduration=134.130745472 podStartE2EDuration="2m14.130745472s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.092060328 +0000 UTC m=+155.448572447" watchObservedRunningTime="2025-11-25 21:53:37.130745472 +0000 UTC m=+155.487257591" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.175192 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.175554 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.675539404 +0000 UTC m=+156.032051523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.277557 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.278171 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.778148146 +0000 UTC m=+156.134660265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.292543 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-gdm7n" podStartSLOduration=133.292493071 podStartE2EDuration="2m13.292493071s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.26044358 +0000 UTC m=+155.616955699" watchObservedRunningTime="2025-11-25 21:53:37.292493071 +0000 UTC m=+155.649005190" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.294150 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-9dxf2" podStartSLOduration=133.294140456 podStartE2EDuration="2m13.294140456s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.287316018 +0000 UTC m=+155.643828137" watchObservedRunningTime="2025-11-25 21:53:37.294140456 +0000 UTC m=+155.650652575" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.383470 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.383965 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.883950546 +0000 UTC m=+156.240462665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.449044 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-vqmcg" podStartSLOduration=133.449020325 podStartE2EDuration="2m13.449020325s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.448131431 +0000 UTC m=+155.804643550" watchObservedRunningTime="2025-11-25 21:53:37.449020325 +0000 UTC m=+155.805532444" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.484971 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.485215 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.98518466 +0000 UTC m=+156.341696789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.485364 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.485724 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:37.985709345 +0000 UTC m=+156.342221464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.489654 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" podStartSLOduration=133.489637382 podStartE2EDuration="2m13.489637382s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.488680956 +0000 UTC m=+155.845193085" watchObservedRunningTime="2025-11-25 21:53:37.489637382 +0000 UTC m=+155.846149501" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.572633 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tqm7k" podStartSLOduration=133.572612875 podStartE2EDuration="2m13.572612875s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.533555511 +0000 UTC m=+155.890067630" watchObservedRunningTime="2025-11-25 21:53:37.572612875 +0000 UTC m=+155.929124994" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.573446 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" podStartSLOduration=133.573438687 podStartE2EDuration="2m13.573438687s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.570158537 +0000 UTC m=+155.926670666" watchObservedRunningTime="2025-11-25 21:53:37.573438687 +0000 UTC m=+155.929950806" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.586327 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.586772 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.086751674 +0000 UTC m=+156.443263793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.623740 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g78xj" podStartSLOduration=134.62372466 podStartE2EDuration="2m14.62372466s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.610304351 +0000 UTC m=+155.966816470" watchObservedRunningTime="2025-11-25 21:53:37.62372466 +0000 UTC m=+155.980236779" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.662067 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-fzvcr" podStartSLOduration=133.662049985 podStartE2EDuration="2m13.662049985s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.659548086 +0000 UTC m=+156.016060205" watchObservedRunningTime="2025-11-25 21:53:37.662049985 +0000 UTC m=+156.018562104" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.687314 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.687603 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.187592617 +0000 UTC m=+156.544104736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.700240 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" event={"ID":"c35b8c1e-cb44-4e0b-9732-0132cdb4e10a","Type":"ContainerStarted","Data":"28cf15b7a03c7daeec51c94dad1bf6c87071aad484839c933ea9afd676a725e4"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.702561 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" event={"ID":"8385c8d3-791b-48c6-9173-84b036a6f294","Type":"ContainerStarted","Data":"fccdc6e8a9001dcaae6e2e1fac9b7228e8ad6fb86fa476837f557ef284098d58"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.704232 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" event={"ID":"640d1c70-7a52-4f0d-a4df-a65d4b247acb","Type":"ContainerStarted","Data":"e2422d673bcec337cd6220f18cec591382f38687cb8bf9dd171bbebe010e84aa"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.704843 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.705703 4931 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hpqqz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.705738 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" podUID="640d1c70-7a52-4f0d-a4df-a65d4b247acb" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.706360 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" event={"ID":"35c9fc97-1683-4083-9090-d1cfcc0a605e","Type":"ContainerStarted","Data":"5e054917c2d928c9813e51137a2a4b0e7d82ada560d89b5e7d76c79ca453f025"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.706383 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" event={"ID":"35c9fc97-1683-4083-9090-d1cfcc0a605e","Type":"ContainerStarted","Data":"eb71df6c1f83c9ef1fe5e69c806a309db6d112b0ea88c4ac6fcadbc255a58873"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.707623 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" event={"ID":"bf866554-0f21-4262-9b8e-32d9c0964488","Type":"ContainerStarted","Data":"b12268119cbe2d57e43972e6270104aa4490c7804bb1592170a0fdd30cb60eb5"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.708134 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.708632 4931 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gkwhd container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.708659 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" podUID="bf866554-0f21-4262-9b8e-32d9c0964488" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.711353 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" event={"ID":"f421de15-2d9f-46ef-82da-414130bdf02e","Type":"ContainerStarted","Data":"437229d5cf46e8567d8a005272eccbb5fe52ebc9b26f1b49956b4108da2b976e"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.711396 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" event={"ID":"f421de15-2d9f-46ef-82da-414130bdf02e","Type":"ContainerStarted","Data":"1a20510389a29514c1c0ba60923cb697593bfb85f98697a14c940ab987043751"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.721590 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" event={"ID":"0211b0cb-4e13-45ce-aab7-a0e6848c2bef","Type":"ContainerStarted","Data":"041f59ee942a37acf34ec735a40405db6867ab049ddc93119e6fb8795ecbac64"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.721648 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" event={"ID":"0211b0cb-4e13-45ce-aab7-a0e6848c2bef","Type":"ContainerStarted","Data":"4ceb03f7cac48fc88e92b93e69f1e0bc0a3bd8ca38b61b55328efdb447065a7c"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.736232 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pr9hq" event={"ID":"fddf1692-d58b-4d3e-88c3-995e2c606b3b","Type":"ContainerStarted","Data":"aa43443168ac026c4695c6dc93e8c338a774db18499cc94c8964dcd8e975e88d"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.736288 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pr9hq" event={"ID":"fddf1692-d58b-4d3e-88c3-995e2c606b3b","Type":"ContainerStarted","Data":"346ba04c441250a871894a3426b720bcd4b538fb36cae44dcf0041b77889d69b"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.740679 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qww9c" podStartSLOduration=133.740661856 podStartE2EDuration="2m13.740661856s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.715367341 +0000 UTC m=+156.071879460" watchObservedRunningTime="2025-11-25 21:53:37.740661856 +0000 UTC m=+156.097173975" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.742712 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" podStartSLOduration=133.742704142 podStartE2EDuration="2m13.742704142s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.740460311 +0000 UTC m=+156.096972440" watchObservedRunningTime="2025-11-25 21:53:37.742704142 +0000 UTC m=+156.099216261" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.755413 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" event={"ID":"c2a5d14a-4296-4064-8267-cf447ce6e41f","Type":"ContainerStarted","Data":"524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.755461 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" event={"ID":"c2a5d14a-4296-4064-8267-cf447ce6e41f","Type":"ContainerStarted","Data":"8d96b9e65d869ba2abccd66f54d696c92c730a7d4b08a98fda8cc2976eacaf3a"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.755945 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.756625 4931 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hsz8c container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.756660 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" podUID="c2a5d14a-4296-4064-8267-cf447ce6e41f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.758218 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" event={"ID":"8c3c979d-1a74-40a6-b24b-572c0ab97f63","Type":"ContainerStarted","Data":"67a2df1bb68dd890c45c5a245960176b34a168c710c3a6d88c35fd1f61219cc5"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.758264 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" event={"ID":"8c3c979d-1a74-40a6-b24b-572c0ab97f63","Type":"ContainerStarted","Data":"486950313a37c860c8ee1a0b7a44e1e8bdebc3bbf58ca6f258eeb25d1789c38f"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.770021 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" event={"ID":"b364e7c6-bd4d-4d58-b348-c700933f4d47","Type":"ContainerStarted","Data":"d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.770045 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" event={"ID":"b364e7c6-bd4d-4d58-b348-c700933f4d47","Type":"ContainerStarted","Data":"29b7e4aa9bdaf8c6a4bf852ebcd52bcbc084001506fc1ca66cf23c97c3d2472c"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.776930 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" podStartSLOduration=133.776911683 podStartE2EDuration="2m13.776911683s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.775245467 +0000 UTC m=+156.131757596" watchObservedRunningTime="2025-11-25 21:53:37.776911683 +0000 UTC m=+156.133423802" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.784244 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" event={"ID":"9fad68f4-2997-42c8-aeb2-937d220f7345","Type":"ContainerStarted","Data":"6ba931d873f61659555d7165ae155c04b185014ae7be984e5e509329a5bff75e"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.784301 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" event={"ID":"9fad68f4-2997-42c8-aeb2-937d220f7345","Type":"ContainerStarted","Data":"b448611017a3d2b8ad5d34683d979fedf8a602676e98cd10c1ce6a16451370e0"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.787951 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.789650 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.289630713 +0000 UTC m=+156.646142832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.797679 4931 generic.go:334] "Generic (PLEG): container finished" podID="2b04f34b-bb66-4247-92e4-c447085d9144" containerID="7d4d16adbef6833081ef05098d7f45e5f1b817d8f1af47ff4b322f480b491429" exitCode=0 Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.797749 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" event={"ID":"2b04f34b-bb66-4247-92e4-c447085d9144","Type":"ContainerDied","Data":"7d4d16adbef6833081ef05098d7f45e5f1b817d8f1af47ff4b322f480b491429"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.801957 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-d7z6d" event={"ID":"e46e33eb-d1b7-4f60-8911-23a2227776de","Type":"ContainerStarted","Data":"b91b2c9353458b03339f3c35055f5df4c35ed01402ab17c662f2d1c93682c4b6"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.813360 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-gkbbs" podStartSLOduration=133.813344275 podStartE2EDuration="2m13.813344275s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.811636628 +0000 UTC m=+156.168148757" watchObservedRunningTime="2025-11-25 21:53:37.813344275 +0000 UTC m=+156.169856394" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.847659 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" event={"ID":"3151332c-78d1-439b-bf22-25e60ed60dd9","Type":"ContainerStarted","Data":"601fb6efa9ee7280e5e7f26d5fba761fb72ab8b35efa3b320904cecdcbb737eb"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.847709 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" event={"ID":"3151332c-78d1-439b-bf22-25e60ed60dd9","Type":"ContainerStarted","Data":"72deddde654101af4340fe59291f5dd0f4f6c5b3ace89852d5837e28c406acd2"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.853790 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" event={"ID":"396991bf-e7c0-448e-a2fc-a626f0bd43b6","Type":"ContainerStarted","Data":"c4afa9ac6d12a25ffa7da1a39773a53f1afeff4542d39a4180661c3c63094f51"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.856511 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" podStartSLOduration=133.856489911 podStartE2EDuration="2m13.856489911s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.854438715 +0000 UTC m=+156.210950834" watchObservedRunningTime="2025-11-25 21:53:37.856489911 +0000 UTC m=+156.213002030" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.862086 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" event={"ID":"f3477fcf-270d-4b95-a581-3cfe8770cd0d","Type":"ContainerStarted","Data":"0946f62ccfb83923dd2ae1bea04e819fc1649bf389502a67619173772e58b0ef"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.862139 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" event={"ID":"f3477fcf-270d-4b95-a581-3cfe8770cd0d","Type":"ContainerStarted","Data":"d760ebea0df9a1f2a8a8f15cb557cdd1d843ac3481edb6ce063ef304f73edacb"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.877801 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" event={"ID":"9d1e5baf-648e-4327-afc5-7a2835d7fb21","Type":"ContainerStarted","Data":"0c1c4c1c51fa61fd57f0c0999558eef6c3535415ce1117a85a987784bc3acc1a"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.882535 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hxhhw" event={"ID":"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4","Type":"ContainerStarted","Data":"4685ca960d863cc2639d4b4320a7dbdc3847f7f4c4ea36f30b4043831f1016b0"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.882580 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hxhhw" event={"ID":"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4","Type":"ContainerStarted","Data":"6946990fb5b0894e530dd10f9d01c910df51e13b2eafc30566ed5b0be304eee8"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.884354 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" event={"ID":"54f4b8f7-8d9c-4bef-af7f-cb2335639e50","Type":"ContainerStarted","Data":"0c092add9fcb05772aa75719aa34fdbaf9e7b8ca408819a53b039acdf9471f6b"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.889628 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.892926 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.392869152 +0000 UTC m=+156.749381271 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.903003 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rwnnf" podStartSLOduration=133.902986891 podStartE2EDuration="2m13.902986891s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.885848769 +0000 UTC m=+156.242360908" watchObservedRunningTime="2025-11-25 21:53:37.902986891 +0000 UTC m=+156.259499000" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.903315 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" event={"ID":"842021b8-6f3c-4eab-bcba-0b27f6736e9e","Type":"ContainerStarted","Data":"c71bda15c6053b9863e4ad19eff0c316dc599981949ac534f81874102b30cc3f"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.903800 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" event={"ID":"842021b8-6f3c-4eab-bcba-0b27f6736e9e","Type":"ContainerStarted","Data":"d3047cb1968e9c75bc8f474d32ea059522b5d5a331f1d307a1bd55fc6aa1ac3f"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.905530 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" event={"ID":"6110aafd-c076-4fa4-a0b6-f5f434b9cfea","Type":"ContainerStarted","Data":"01defdf08b94e47ce4a91449f5e4d86b079b3d85db5c52e623a346875de99b80"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.908488 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m" event={"ID":"b68a4880-028d-4379-ac82-67d9a9f6a1d3","Type":"ContainerStarted","Data":"0aaf3636342fc317ad8467194195efe823ec782df45b40e563f53d6fd1802e6b"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.915550 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" event={"ID":"c036f4c8-f544-4f0d-b1c6-51944420430c","Type":"ContainerStarted","Data":"a525778e298f6bc14bc2e734f045a374ff7899c5bd6f4b8d16546fe8bae5cf2d"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.915575 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" event={"ID":"c036f4c8-f544-4f0d-b1c6-51944420430c","Type":"ContainerStarted","Data":"ab001ca2bf3618ebc8f97756c6e5cfd925322969822205db12584758b31c7066"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.917440 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" event={"ID":"5af8ff72-8043-4f10-8357-10ce2ff5123b","Type":"ContainerStarted","Data":"9bba3434ea75eef192c53403170324516bb60395cd3d42767e3ad1cdabca97ea"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.917469 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" event={"ID":"5af8ff72-8043-4f10-8357-10ce2ff5123b","Type":"ContainerStarted","Data":"d8480ebcc51cf38697eb696856cc4e0a1f5f6682a54971f25df5e44c93af19cf"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.918963 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" event={"ID":"53a6c85d-8623-46b6-9d53-860f1ff80f00","Type":"ContainerStarted","Data":"581b9740a710c4aff4fdb3c733df718af92e87f021a5eb7340eaad6d298b3378"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.919415 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.926035 4931 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2gvlz container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.926099 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" podUID="53a6c85d-8623-46b6-9d53-860f1ff80f00" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.926106 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" event={"ID":"1ab99d3e-35d5-4a34-b195-fc53434b5168","Type":"ContainerStarted","Data":"ee72799a64a45a1822c799dc1fb9651ecf8939bf1664c9b1df0773380815a73c"} Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.927308 4931 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-454st container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.927362 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" podUID="fada681c-6ade-49e4-b324-1985853bc61e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.927491 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.929919 4931 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kvz7t container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.929980 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" podUID="74cbb6cf-4717-439c-b712-67718c154742" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.929977 4931 patch_prober.go:28] interesting pod/console-operator-58897d9998-vqmcg container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.930032 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-vqmcg" podUID="b4dd52d7-4597-4159-a728-012cd2180b35" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.940218 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-6vpzc" podStartSLOduration=133.940194603 podStartE2EDuration="2m13.940194603s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.935420973 +0000 UTC m=+156.291933092" watchObservedRunningTime="2025-11-25 21:53:37.940194603 +0000 UTC m=+156.296706722" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.984465 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:37 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:37 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:37 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.984519 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.990785 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:37 crc kubenswrapper[4931]: E1125 21:53:37.990966 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.490943779 +0000 UTC m=+156.847455898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:37 crc kubenswrapper[4931]: I1125 21:53:37.991274 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.002551 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.502533208 +0000 UTC m=+156.859045327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.013510 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-pr9hq" podStartSLOduration=7.013478939 podStartE2EDuration="7.013478939s" podCreationTimestamp="2025-11-25 21:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:37.979262858 +0000 UTC m=+156.335774977" watchObservedRunningTime="2025-11-25 21:53:38.013478939 +0000 UTC m=+156.369991058" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.024180 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-d7z6d" podStartSLOduration=7.024157163 podStartE2EDuration="7.024157163s" podCreationTimestamp="2025-11-25 21:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.011069643 +0000 UTC m=+156.367581762" watchObservedRunningTime="2025-11-25 21:53:38.024157163 +0000 UTC m=+156.380669282" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.098344 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.098711 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.598694562 +0000 UTC m=+156.955206681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.104938 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" podStartSLOduration=134.104925644 podStartE2EDuration="2m14.104925644s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.103925926 +0000 UTC m=+156.460438045" watchObservedRunningTime="2025-11-25 21:53:38.104925644 +0000 UTC m=+156.461437763" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.138700 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" podStartSLOduration=134.138681262 podStartE2EDuration="2m14.138681262s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.132369298 +0000 UTC m=+156.488881437" watchObservedRunningTime="2025-11-25 21:53:38.138681262 +0000 UTC m=+156.495193371" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.182574 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vgppb" podStartSLOduration=134.182556019 podStartE2EDuration="2m14.182556019s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.177365627 +0000 UTC m=+156.533877746" watchObservedRunningTime="2025-11-25 21:53:38.182556019 +0000 UTC m=+156.539068138" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.199551 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.199849 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.699837145 +0000 UTC m=+157.056349264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.213364 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" podStartSLOduration=134.213351046 podStartE2EDuration="2m14.213351046s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.211348161 +0000 UTC m=+156.567860280" watchObservedRunningTime="2025-11-25 21:53:38.213351046 +0000 UTC m=+156.569863165" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.252378 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4pgh" podStartSLOduration=134.252360659 podStartE2EDuration="2m14.252360659s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.250085366 +0000 UTC m=+156.606597485" watchObservedRunningTime="2025-11-25 21:53:38.252360659 +0000 UTC m=+156.608872778" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.288707 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w6tf9" podStartSLOduration=134.288691308 podStartE2EDuration="2m14.288691308s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.287255998 +0000 UTC m=+156.643768117" watchObservedRunningTime="2025-11-25 21:53:38.288691308 +0000 UTC m=+156.645203427" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.300885 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.301122 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.801093489 +0000 UTC m=+157.157605638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.301387 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.301723 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.801707825 +0000 UTC m=+157.158219944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.333005 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-f5nxk" podStartSLOduration=135.332986976 podStartE2EDuration="2m15.332986976s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.331028182 +0000 UTC m=+156.687540371" watchObservedRunningTime="2025-11-25 21:53:38.332986976 +0000 UTC m=+156.689499095" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.372609 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc87m" podStartSLOduration=134.372592915 podStartE2EDuration="2m14.372592915s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.371916617 +0000 UTC m=+156.728428736" watchObservedRunningTime="2025-11-25 21:53:38.372592915 +0000 UTC m=+156.729105034" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.402410 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.402743 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:38.902720054 +0000 UTC m=+157.259232173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.409555 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-gmjcj" podStartSLOduration=134.409536591 podStartE2EDuration="2m14.409536591s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.407255559 +0000 UTC m=+156.763767678" watchObservedRunningTime="2025-11-25 21:53:38.409536591 +0000 UTC m=+156.766048710" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.503738 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.505487 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.005467439 +0000 UTC m=+157.361979618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.605120 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.605290 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.105262874 +0000 UTC m=+157.461774993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.605789 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.606154 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.106145538 +0000 UTC m=+157.462657647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.706660 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.706862 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.206836998 +0000 UTC m=+157.563349117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.707174 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.707482 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.207469895 +0000 UTC m=+157.563982014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.726296 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.726356 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.727979 4931 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-8jz7m container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.12:8443/livez\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.728072 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" podUID="9d1e5baf-648e-4327-afc5-7a2835d7fb21" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.12:8443/livez\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.808192 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.808369 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.308344979 +0000 UTC m=+157.664857098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.808546 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.808864 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.308857273 +0000 UTC m=+157.665369392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.843629 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.843691 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.909337 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.909589 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.409541573 +0000 UTC m=+157.766053692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.909620 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:38 crc kubenswrapper[4931]: E1125 21:53:38.909984 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.409972684 +0000 UTC m=+157.766484863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.933046 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hxhhw" event={"ID":"45bb8c11-1f6e-4f8a-a67d-b2f531211ba4","Type":"ContainerStarted","Data":"18d37eb01d8da86588b33f02950aba02ee913ba27c30c027feadd07998bb22b1"} Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.933110 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.934796 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" event={"ID":"842021b8-6f3c-4eab-bcba-0b27f6736e9e","Type":"ContainerStarted","Data":"aa9144c0717005ced5aff17359c35f9960b3e924137b8c995a89096473358731"} Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.936370 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" event={"ID":"f421de15-2d9f-46ef-82da-414130bdf02e","Type":"ContainerStarted","Data":"d81672a23e2fe7c5d92b45446fa8b75e1f8d398cc96a261940f3a64f18ac2f92"} Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.937940 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" event={"ID":"35c9fc97-1683-4083-9090-d1cfcc0a605e","Type":"ContainerStarted","Data":"c62666c77facb17bc86838a8ecf456d2a1280a59b96fd4a1bfa6f125964b63db"} Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.938014 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.939510 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" event={"ID":"2b04f34b-bb66-4247-92e4-c447085d9144","Type":"ContainerStarted","Data":"22cd957479fe032941f00ed14d8236cb604dd5374e0842dedf9b55bf7b93258a"} Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.939625 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.940990 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" event={"ID":"5af8ff72-8043-4f10-8357-10ce2ff5123b","Type":"ContainerStarted","Data":"405ca308bef817ff1eb27eb0d42d3172794a270f28a8768522c2f8f9c3b8bc17"} Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941434 4931 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hpqqz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941483 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" podUID="640d1c70-7a52-4f0d-a4df-a65d4b247acb" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941434 4931 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-454st container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941754 4931 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gkwhd container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941771 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" podUID="fada681c-6ade-49e4-b324-1985853bc61e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941792 4931 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hsz8c container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941789 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" podUID="bf866554-0f21-4262-9b8e-32d9c0964488" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941839 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" podUID="c2a5d14a-4296-4064-8267-cf447ce6e41f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941964 4931 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2gvlz container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.941982 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" podUID="53a6c85d-8623-46b6-9d53-860f1ff80f00" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.942179 4931 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kvz7t container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.942207 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" podUID="74cbb6cf-4717-439c-b712-67718c154742" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.965952 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wnjvs" podStartSLOduration=134.965935123 podStartE2EDuration="2m14.965935123s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.471962848 +0000 UTC m=+156.828474957" watchObservedRunningTime="2025-11-25 21:53:38.965935123 +0000 UTC m=+157.322447232" Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.967953 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:38 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:38 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:38 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:38 crc kubenswrapper[4931]: I1125 21:53:38.967994 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.010919 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.013121 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.513106141 +0000 UTC m=+157.869618250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.049186 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-hxhhw" podStartSLOduration=8.049171182 podStartE2EDuration="8.049171182s" podCreationTimestamp="2025-11-25 21:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:38.96765023 +0000 UTC m=+157.324162349" watchObservedRunningTime="2025-11-25 21:53:39.049171182 +0000 UTC m=+157.405683301" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.105923 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kxjw7" podStartSLOduration=135.105903333 podStartE2EDuration="2m15.105903333s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:39.066000245 +0000 UTC m=+157.422512364" watchObservedRunningTime="2025-11-25 21:53:39.105903333 +0000 UTC m=+157.462415452" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.113248 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.113620 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.613584073 +0000 UTC m=+157.970096192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.133753 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" podStartSLOduration=135.133739658 podStartE2EDuration="2m15.133739658s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:39.132542225 +0000 UTC m=+157.489054374" watchObservedRunningTime="2025-11-25 21:53:39.133739658 +0000 UTC m=+157.490251777" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.135043 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" podStartSLOduration=136.135028614 podStartE2EDuration="2m16.135028614s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:39.103607449 +0000 UTC m=+157.460119568" watchObservedRunningTime="2025-11-25 21:53:39.135028614 +0000 UTC m=+157.491540733" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.159258 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-s2td2" podStartSLOduration=135.159208929 podStartE2EDuration="2m15.159208929s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:39.156027111 +0000 UTC m=+157.512539230" watchObservedRunningTime="2025-11-25 21:53:39.159208929 +0000 UTC m=+157.515721048" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.182486 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5dlk4" podStartSLOduration=135.182454177 podStartE2EDuration="2m15.182454177s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:39.180861754 +0000 UTC m=+157.537373873" watchObservedRunningTime="2025-11-25 21:53:39.182454177 +0000 UTC m=+157.538966296" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.214036 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.214405 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.714389856 +0000 UTC m=+158.070901965 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.263233 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" podStartSLOduration=136.263217529 podStartE2EDuration="2m16.263217529s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:39.262252412 +0000 UTC m=+157.618764521" watchObservedRunningTime="2025-11-25 21:53:39.263217529 +0000 UTC m=+157.619729648" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.265117 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tp4r9" podStartSLOduration=135.265110911 podStartE2EDuration="2m15.265110911s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:39.221691106 +0000 UTC m=+157.578203225" watchObservedRunningTime="2025-11-25 21:53:39.265110911 +0000 UTC m=+157.621623030" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.291309 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" podStartSLOduration=136.291294601 podStartE2EDuration="2m16.291294601s" podCreationTimestamp="2025-11-25 21:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:39.28872933 +0000 UTC m=+157.645241449" watchObservedRunningTime="2025-11-25 21:53:39.291294601 +0000 UTC m=+157.647806720" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.315419 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.315749 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.815737503 +0000 UTC m=+158.172249622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.416061 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.416230 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.916205286 +0000 UTC m=+158.272717405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.416344 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.416617 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:39.916606817 +0000 UTC m=+158.273118936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.517430 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.518336 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.018010026 +0000 UTC m=+158.374522145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.618619 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.618963 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.118946732 +0000 UTC m=+158.475458851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.719830 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.719953 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.219934629 +0000 UTC m=+158.576446748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.720197 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.720506 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.220499855 +0000 UTC m=+158.577011974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.821941 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.822084 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.322063817 +0000 UTC m=+158.678575936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.822206 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.822573 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.322562222 +0000 UTC m=+158.679074341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.923783 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.924017 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.423987171 +0000 UTC m=+158.780499300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.924188 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:39 crc kubenswrapper[4931]: E1125 21:53:39.924607 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.424596658 +0000 UTC m=+158.781108867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.948313 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" event={"ID":"1da80646-0675-4f6c-ac13-3e19f44ce6a3","Type":"ContainerStarted","Data":"b6131398998c03478efac8200d1905db14950c1708eb07b0e3943027d3df0a17"} Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.948985 4931 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hsz8c container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.949021 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" podUID="c2a5d14a-4296-4064-8267-cf447ce6e41f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.949444 4931 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hpqqz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.949483 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" podUID="640d1c70-7a52-4f0d-a4df-a65d4b247acb" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.953496 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gkwhd" Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.965982 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:39 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:39 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:39 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:39 crc kubenswrapper[4931]: I1125 21:53:39.966046 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.025582 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.025760 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.525719179 +0000 UTC m=+158.882231298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.027039 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.027640 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.527628271 +0000 UTC m=+158.884140390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.129013 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.129150 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.629129373 +0000 UTC m=+158.985641492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.129550 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.129917 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.629906064 +0000 UTC m=+158.986418183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.230382 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.230767 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.730748997 +0000 UTC m=+159.087261116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.332307 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.332640 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.832623899 +0000 UTC m=+159.189136018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.433307 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.433435 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.933416631 +0000 UTC m=+159.289928750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.433529 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.434040 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:40.934030267 +0000 UTC m=+159.290542376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.534503 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.535161 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.035138529 +0000 UTC m=+159.391650658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.636598 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.637040 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.1370203 +0000 UTC m=+159.493532419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.737355 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.737496 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.237468373 +0000 UTC m=+159.593980492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.737665 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.738005 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.237994667 +0000 UTC m=+159.594506786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.839057 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.839213 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.339191501 +0000 UTC m=+159.695703620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.839720 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.840103 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.340092695 +0000 UTC m=+159.696604814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.860096 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2gvlz" Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.880634 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.909512 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.945656 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:40 crc kubenswrapper[4931]: E1125 21:53:40.946662 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.446647295 +0000 UTC m=+159.803159404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.966054 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:40 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:40 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:40 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:40 crc kubenswrapper[4931]: I1125 21:53:40.966100 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.047840 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.049208 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.549195666 +0000 UTC m=+159.905707785 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.102730 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qmg94"] Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.103996 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.110397 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.129116 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qmg94"] Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.149245 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.149446 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.649420082 +0000 UTC m=+160.005932201 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.149551 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-utilities\") pod \"certified-operators-qmg94\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.149598 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6kks\" (UniqueName: \"kubernetes.io/projected/ff6d71fd-060b-411c-a576-2651944470e3-kube-api-access-f6kks\") pod \"certified-operators-qmg94\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.149693 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-catalog-content\") pod \"certified-operators-qmg94\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.149794 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.150099 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.650092041 +0000 UTC m=+160.006604160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.251126 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.251335 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6kks\" (UniqueName: \"kubernetes.io/projected/ff6d71fd-060b-411c-a576-2651944470e3-kube-api-access-f6kks\") pod \"certified-operators-qmg94\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.251392 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-catalog-content\") pod \"certified-operators-qmg94\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.251443 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-utilities\") pod \"certified-operators-qmg94\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.251807 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-utilities\") pod \"certified-operators-qmg94\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.251892 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.751861339 +0000 UTC m=+160.108373448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.252339 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-catalog-content\") pod \"certified-operators-qmg94\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.280645 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6kks\" (UniqueName: \"kubernetes.io/projected/ff6d71fd-060b-411c-a576-2651944470e3-kube-api-access-f6kks\") pod \"certified-operators-qmg94\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.352833 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.353189 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.853177916 +0000 UTC m=+160.209690035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.421086 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.454105 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.454399 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:41.954384789 +0000 UTC m=+160.310896908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.502097 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fn56g"] Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.504543 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.519300 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fn56g"] Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.555103 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-catalog-content\") pod \"certified-operators-fn56g\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.555178 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-utilities\") pod \"certified-operators-fn56g\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.555230 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.555314 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phgnm\" (UniqueName: \"kubernetes.io/projected/16d28da7-dc24-4c86-914b-1bf11af8c9fd-kube-api-access-phgnm\") pod \"certified-operators-fn56g\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.555481 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.055469809 +0000 UTC m=+160.411981928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.661456 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.661648 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.161619848 +0000 UTC m=+160.518131967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.661763 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-catalog-content\") pod \"certified-operators-fn56g\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.662591 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-catalog-content\") pod \"certified-operators-fn56g\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.662617 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-utilities\") pod \"certified-operators-fn56g\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.661869 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-utilities\") pod \"certified-operators-fn56g\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.662712 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.663036 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.163027927 +0000 UTC m=+160.519540046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.663173 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phgnm\" (UniqueName: \"kubernetes.io/projected/16d28da7-dc24-4c86-914b-1bf11af8c9fd-kube-api-access-phgnm\") pod \"certified-operators-fn56g\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.688220 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-glk9x"] Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.689132 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.696173 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.710492 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-glk9x"] Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.716153 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phgnm\" (UniqueName: \"kubernetes.io/projected/16d28da7-dc24-4c86-914b-1bf11af8c9fd-kube-api-access-phgnm\") pod \"certified-operators-fn56g\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.765090 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.765606 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-utilities\") pod \"community-operators-glk9x\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.765653 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-catalog-content\") pod \"community-operators-glk9x\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.765676 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwv4c\" (UniqueName: \"kubernetes.io/projected/abaedc86-626a-4a22-8f88-1738e338e27f-kube-api-access-nwv4c\") pod \"community-operators-glk9x\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.765836 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.265821394 +0000 UTC m=+160.622333513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.822136 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.866853 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-catalog-content\") pod \"community-operators-glk9x\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.866903 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwv4c\" (UniqueName: \"kubernetes.io/projected/abaedc86-626a-4a22-8f88-1738e338e27f-kube-api-access-nwv4c\") pod \"community-operators-glk9x\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.866959 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.867002 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-utilities\") pod \"community-operators-glk9x\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.867409 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-utilities\") pod \"community-operators-glk9x\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.867789 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-catalog-content\") pod \"community-operators-glk9x\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.868357 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.368344284 +0000 UTC m=+160.724856403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.894642 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bnlql"] Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.895537 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.927168 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bnlql"] Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.934222 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwv4c\" (UniqueName: \"kubernetes.io/projected/abaedc86-626a-4a22-8f88-1738e338e27f-kube-api-access-nwv4c\") pod \"community-operators-glk9x\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.944932 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qmg94"] Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.966413 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:41 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:41 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:41 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.966498 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.967800 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.968023 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-utilities\") pod \"community-operators-bnlql\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.968147 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-catalog-content\") pod \"community-operators-bnlql\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:41 crc kubenswrapper[4931]: I1125 21:53:41.968227 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bnt7\" (UniqueName: \"kubernetes.io/projected/7d2bf749-d61d-4049-a314-dacdf107bdb5-kube-api-access-7bnt7\") pod \"community-operators-bnlql\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:41 crc kubenswrapper[4931]: E1125 21:53:41.968371 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.468336783 +0000 UTC m=+160.824848912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.038097 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.072199 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.072262 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-catalog-content\") pod \"community-operators-bnlql\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.072316 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bnt7\" (UniqueName: \"kubernetes.io/projected/7d2bf749-d61d-4049-a314-dacdf107bdb5-kube-api-access-7bnt7\") pod \"community-operators-bnlql\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.072343 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-utilities\") pod \"community-operators-bnlql\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.072754 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-utilities\") pod \"community-operators-bnlql\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.073067 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.573056233 +0000 UTC m=+160.929568352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.073962 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-catalog-content\") pod \"community-operators-bnlql\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.104622 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bnt7\" (UniqueName: \"kubernetes.io/projected/7d2bf749-d61d-4049-a314-dacdf107bdb5-kube-api-access-7bnt7\") pod \"community-operators-bnlql\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.173450 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.173604 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.673579758 +0000 UTC m=+161.030091877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.173998 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.174278 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.674266106 +0000 UTC m=+161.030778215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.253158 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.274849 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.275001 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.774984556 +0000 UTC m=+161.131496675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.275423 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.275881 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.77583956 +0000 UTC m=+161.132351679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.377505 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.377816 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.877802085 +0000 UTC m=+161.234314204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.441651 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.442461 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.445226 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.445508 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.478637 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60e97735-4431-4143-ba6d-ec246a23722b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"60e97735-4431-4143-ba6d-ec246a23722b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.478682 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60e97735-4431-4143-ba6d-ec246a23722b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"60e97735-4431-4143-ba6d-ec246a23722b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.478730 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.478999 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:42.978989347 +0000 UTC m=+161.335501466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.480741 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.485172 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-glk9x"] Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.508356 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fn56g"] Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.579116 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.579282 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60e97735-4431-4143-ba6d-ec246a23722b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"60e97735-4431-4143-ba6d-ec246a23722b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.579337 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60e97735-4431-4143-ba6d-ec246a23722b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"60e97735-4431-4143-ba6d-ec246a23722b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.579389 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.079371248 +0000 UTC m=+161.435883377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.579458 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.579744 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60e97735-4431-4143-ba6d-ec246a23722b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"60e97735-4431-4143-ba6d-ec246a23722b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.579767 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.079755098 +0000 UTC m=+161.436267227 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.591233 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bnlql"] Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.604231 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60e97735-4431-4143-ba6d-ec246a23722b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"60e97735-4431-4143-ba6d-ec246a23722b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.680976 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.681112 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.181090485 +0000 UTC m=+161.537602614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.681564 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.682187 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.182154054 +0000 UTC m=+161.538666173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.763701 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.782534 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.782840 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.282821803 +0000 UTC m=+161.639333942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.884358 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.884650 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.384636993 +0000 UTC m=+161.741149112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.967122 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:42 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:42 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:42 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.967388 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:42 crc kubenswrapper[4931]: I1125 21:53:42.987772 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:42 crc kubenswrapper[4931]: E1125 21:53:42.988227 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.488206962 +0000 UTC m=+161.844719081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.012483 4931 generic.go:334] "Generic (PLEG): container finished" podID="abaedc86-626a-4a22-8f88-1738e338e27f" containerID="5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f" exitCode=0 Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.012547 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glk9x" event={"ID":"abaedc86-626a-4a22-8f88-1738e338e27f","Type":"ContainerDied","Data":"5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f"} Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.012573 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glk9x" event={"ID":"abaedc86-626a-4a22-8f88-1738e338e27f","Type":"ContainerStarted","Data":"7cb61b740189b3a49b4e2ba7d6c0426b0077d76de421eeb4bb55258fdca9b3e0"} Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.016285 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.017694 4931 generic.go:334] "Generic (PLEG): container finished" podID="ff6d71fd-060b-411c-a576-2651944470e3" containerID="4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96" exitCode=0 Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.017743 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmg94" event={"ID":"ff6d71fd-060b-411c-a576-2651944470e3","Type":"ContainerDied","Data":"4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96"} Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.017767 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmg94" event={"ID":"ff6d71fd-060b-411c-a576-2651944470e3","Type":"ContainerStarted","Data":"c30008e5c34bcdf9485f3cd53954e4b9e308486aed3a43d0bb9182f9d46f11e7"} Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.034194 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" event={"ID":"1da80646-0675-4f6c-ac13-3e19f44ce6a3","Type":"ContainerStarted","Data":"c7aa45109ea0a3bb08ef44fe076654b55ec2735ec0452bce5738cb5ef810812c"} Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.060168 4931 generic.go:334] "Generic (PLEG): container finished" podID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerID="2a2fc74dc769b0a35c7f235ffabe2e0935121ddb624c5b9d7fc3fe4bd14bd7f0" exitCode=0 Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.060228 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fn56g" event={"ID":"16d28da7-dc24-4c86-914b-1bf11af8c9fd","Type":"ContainerDied","Data":"2a2fc74dc769b0a35c7f235ffabe2e0935121ddb624c5b9d7fc3fe4bd14bd7f0"} Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.060254 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fn56g" event={"ID":"16d28da7-dc24-4c86-914b-1bf11af8c9fd","Type":"ContainerStarted","Data":"b85a88fd87699a702704b283f1f6addc357034da91c932b0cb589ce0b4f39f62"} Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.062655 4931 generic.go:334] "Generic (PLEG): container finished" podID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerID="fefd4f318f057ab6277310db0ac2b7c615321f175488b48fc7bec511d778a3bf" exitCode=0 Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.062690 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnlql" event={"ID":"7d2bf749-d61d-4049-a314-dacdf107bdb5","Type":"ContainerDied","Data":"fefd4f318f057ab6277310db0ac2b7c615321f175488b48fc7bec511d778a3bf"} Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.062714 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnlql" event={"ID":"7d2bf749-d61d-4049-a314-dacdf107bdb5","Type":"ContainerStarted","Data":"3ea90d2fe461cd291d86ef50f25fcbd935cd22b210d85baaab5105136fb72cd0"} Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.083105 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.093444 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.094772 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.594705791 +0000 UTC m=+161.951217910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.196348 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.196521 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.696494099 +0000 UTC m=+162.053006218 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.196674 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.196986 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.696973693 +0000 UTC m=+162.053485812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.297326 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.297425 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.797411194 +0000 UTC m=+162.153923303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.297594 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.297971 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.79795572 +0000 UTC m=+162.154467839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.398948 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.399148 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.899113662 +0000 UTC m=+162.255625801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.399208 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.399887 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:43.899856353 +0000 UTC m=+162.256368482 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.499942 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.500580 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.000565122 +0000 UTC m=+162.357077241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.502566 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pc84d"] Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.503799 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.509066 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.525249 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc84d"] Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.601557 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-catalog-content\") pod \"redhat-marketplace-pc84d\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.601615 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-utilities\") pod \"redhat-marketplace-pc84d\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.601714 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.601753 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4mjz\" (UniqueName: \"kubernetes.io/projected/b5d6d339-fbd3-461f-8862-340f1d44c2f8-kube-api-access-v4mjz\") pod \"redhat-marketplace-pc84d\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.602105 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.102091044 +0000 UTC m=+162.458603163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.627484 4931 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.702662 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.702758 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.202740212 +0000 UTC m=+162.559252331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.703243 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.703342 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4mjz\" (UniqueName: \"kubernetes.io/projected/b5d6d339-fbd3-461f-8862-340f1d44c2f8-kube-api-access-v4mjz\") pod \"redhat-marketplace-pc84d\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.703501 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.203493102 +0000 UTC m=+162.560005221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.703499 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-catalog-content\") pod \"redhat-marketplace-pc84d\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.703553 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-utilities\") pod \"redhat-marketplace-pc84d\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.703945 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-utilities\") pod \"redhat-marketplace-pc84d\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.704359 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-catalog-content\") pod \"redhat-marketplace-pc84d\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.720787 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4mjz\" (UniqueName: \"kubernetes.io/projected/b5d6d339-fbd3-461f-8862-340f1d44c2f8-kube-api-access-v4mjz\") pod \"redhat-marketplace-pc84d\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.731242 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.738810 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8jz7m" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.748740 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.748781 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.750460 4931 patch_prober.go:28] interesting pod/console-f9d7485db-2hpxq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.750511 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2hpxq" podUID="69dfd797-b979-42a6-a4d5-d95e48bf7aa2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.767026 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.767243 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.779124 4931 patch_prober.go:28] interesting pod/apiserver-76f77b778f-f8tsr container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]log ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]etcd ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/generic-apiserver-start-informers ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/max-in-flight-filter ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 25 21:53:43 crc kubenswrapper[4931]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/project.openshift.io-projectcache ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/openshift.io-startinformers ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 25 21:53:43 crc kubenswrapper[4931]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 25 21:53:43 crc kubenswrapper[4931]: livez check failed Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.779906 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" podUID="3151332c-78d1-439b-bf22-25e60ed60dd9" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.779681 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.780182 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.780266 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.780305 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.806448 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.806735 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.306720421 +0000 UTC m=+162.663232540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.829323 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.884576 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rj5p7"] Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.885832 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.907505 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:43 crc kubenswrapper[4931]: E1125 21:53:43.909020 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.409002894 +0000 UTC m=+162.765515023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.932501 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rj5p7"] Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.960354 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.963116 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.967075 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:43 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:43 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:43 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:43 crc kubenswrapper[4931]: I1125 21:53:43.967141 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.010750 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.010887 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.510847615 +0000 UTC m=+162.867359734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.011216 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtffm\" (UniqueName: \"kubernetes.io/projected/e82c08a2-5336-42fa-b297-c432ddd981f0-kube-api-access-xtffm\") pod \"redhat-marketplace-rj5p7\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.011283 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-utilities\") pod \"redhat-marketplace-rj5p7\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.011338 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-catalog-content\") pod \"redhat-marketplace-rj5p7\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.011378 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.013020 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.513008915 +0000 UTC m=+162.869521034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.014234 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9wxs5" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.097122 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"60e97735-4431-4143-ba6d-ec246a23722b","Type":"ContainerStarted","Data":"2157b2864740e777e186079bdc4e0060b4b7200710abbff10582b1caaf76290d"} Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.097518 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"60e97735-4431-4143-ba6d-ec246a23722b","Type":"ContainerStarted","Data":"638468860ccb77aa5d45c59744578aa6f0f1b3198ae7000fd826102c19c001e6"} Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.106834 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" event={"ID":"1da80646-0675-4f6c-ac13-3e19f44ce6a3","Type":"ContainerStarted","Data":"937c2487b9e73a2e7d656a654cf2dfa6a3d58b416898e39850e97f8e38f86880"} Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.107057 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" event={"ID":"1da80646-0675-4f6c-ac13-3e19f44ce6a3","Type":"ContainerStarted","Data":"e1cd14d16fd0ac35aa604a35b941f6524ad8a65f1f2b093347412f41a3f38faf"} Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.111852 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.111953 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.611931795 +0000 UTC m=+162.968443914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.112706 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtffm\" (UniqueName: \"kubernetes.io/projected/e82c08a2-5336-42fa-b297-c432ddd981f0-kube-api-access-xtffm\") pod \"redhat-marketplace-rj5p7\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.112779 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-utilities\") pod \"redhat-marketplace-rj5p7\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.112858 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-catalog-content\") pod \"redhat-marketplace-rj5p7\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.112915 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.123132 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-catalog-content\") pod \"redhat-marketplace-rj5p7\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.123734 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-utilities\") pod \"redhat-marketplace-rj5p7\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.124236 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.624220933 +0000 UTC m=+162.980733042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.126767 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc84d"] Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.143811 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtffm\" (UniqueName: \"kubernetes.io/projected/e82c08a2-5336-42fa-b297-c432ddd981f0-kube-api-access-xtffm\") pod \"redhat-marketplace-rj5p7\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.147766 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2c4n7" podStartSLOduration=13.14770697 podStartE2EDuration="13.14770697s" podCreationTimestamp="2025-11-25 21:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:44.142384083 +0000 UTC m=+162.498896202" watchObservedRunningTime="2025-11-25 21:53:44.14770697 +0000 UTC m=+162.504219089" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.207910 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.213642 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.213781 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.713762066 +0000 UTC m=+163.070274185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.213956 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.214223 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.714210908 +0000 UTC m=+163.070723027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.314543 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.314980 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.814965279 +0000 UTC m=+163.171477398 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.315781 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-vqmcg" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.414056 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.415920 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.416340 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:44.916326367 +0000 UTC m=+163.272838486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.481294 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bgfgw"] Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.482593 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.487177 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.491526 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bgfgw"] Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.517392 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.517559 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:45.01753548 +0000 UTC m=+163.374047599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.517823 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-utilities\") pod \"redhat-operators-bgfgw\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.517917 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.517973 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-catalog-content\") pod \"redhat-operators-bgfgw\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.518041 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wgzn\" (UniqueName: \"kubernetes.io/projected/6e742d61-fa3f-494b-aa90-0bd032106c7b-kube-api-access-5wgzn\") pod \"redhat-operators-bgfgw\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.518259 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 21:53:45.018251689 +0000 UTC m=+163.374763808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-84p7l" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.618537 4931 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T21:53:43.627509963Z","Handler":null,"Name":""} Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.619772 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.619896 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wgzn\" (UniqueName: \"kubernetes.io/projected/6e742d61-fa3f-494b-aa90-0bd032106c7b-kube-api-access-5wgzn\") pod \"redhat-operators-bgfgw\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.619982 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-utilities\") pod \"redhat-operators-bgfgw\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.620039 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-catalog-content\") pod \"redhat-operators-bgfgw\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.620408 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-catalog-content\") pod \"redhat-operators-bgfgw\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: E1125 21:53:44.620467 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 21:53:45.12045415 +0000 UTC m=+163.476966269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.621118 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-utilities\") pod \"redhat-operators-bgfgw\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.627916 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rj5p7"] Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.633761 4931 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.633796 4931 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 21:53:44 crc kubenswrapper[4931]: W1125 21:53:44.645929 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode82c08a2_5336_42fa_b297_c432ddd981f0.slice/crio-e3c883a54228a15f29cf4cd262ca8def6f898297e9dfc4fae8ab90bc52c2cfab WatchSource:0}: Error finding container e3c883a54228a15f29cf4cd262ca8def6f898297e9dfc4fae8ab90bc52c2cfab: Status 404 returned error can't find the container with id e3c883a54228a15f29cf4cd262ca8def6f898297e9dfc4fae8ab90bc52c2cfab Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.654520 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wgzn\" (UniqueName: \"kubernetes.io/projected/6e742d61-fa3f-494b-aa90-0bd032106c7b-kube-api-access-5wgzn\") pod \"redhat-operators-bgfgw\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.720821 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.725988 4931 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.726040 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.776729 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-84p7l\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.800399 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.821340 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.830574 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.889139 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5gt24"] Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.890352 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.899938 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5gt24"] Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.901263 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.965510 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:44 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:44 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:44 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:44 crc kubenswrapper[4931]: I1125 21:53:44.965592 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.024179 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-utilities\") pod \"redhat-operators-5gt24\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.024244 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n8ql\" (UniqueName: \"kubernetes.io/projected/1e53eac6-a437-438d-b5ba-6747277367b8-kube-api-access-4n8ql\") pod \"redhat-operators-5gt24\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.024282 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-catalog-content\") pod \"redhat-operators-5gt24\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.065545 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bgfgw"] Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.099763 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.126201 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n8ql\" (UniqueName: \"kubernetes.io/projected/1e53eac6-a437-438d-b5ba-6747277367b8-kube-api-access-4n8ql\") pod \"redhat-operators-5gt24\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.126283 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-catalog-content\") pod \"redhat-operators-5gt24\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.126400 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-utilities\") pod \"redhat-operators-5gt24\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.126857 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-utilities\") pod \"redhat-operators-5gt24\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.127000 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-catalog-content\") pod \"redhat-operators-5gt24\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.133620 4931 generic.go:334] "Generic (PLEG): container finished" podID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerID="eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c" exitCode=0 Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.133691 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc84d" event={"ID":"b5d6d339-fbd3-461f-8862-340f1d44c2f8","Type":"ContainerDied","Data":"eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c"} Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.133723 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc84d" event={"ID":"b5d6d339-fbd3-461f-8862-340f1d44c2f8","Type":"ContainerStarted","Data":"04a46770fe8d445ca9c661ed848953ab3daee65e1fc3f985467aeb65b495c0a9"} Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.152209 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n8ql\" (UniqueName: \"kubernetes.io/projected/1e53eac6-a437-438d-b5ba-6747277367b8-kube-api-access-4n8ql\") pod \"redhat-operators-5gt24\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.155839 4931 generic.go:334] "Generic (PLEG): container finished" podID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerID="df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30" exitCode=0 Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.155995 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rj5p7" event={"ID":"e82c08a2-5336-42fa-b297-c432ddd981f0","Type":"ContainerDied","Data":"df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30"} Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.156020 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rj5p7" event={"ID":"e82c08a2-5336-42fa-b297-c432ddd981f0","Type":"ContainerStarted","Data":"e3c883a54228a15f29cf4cd262ca8def6f898297e9dfc4fae8ab90bc52c2cfab"} Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.174510 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bgfgw" event={"ID":"6e742d61-fa3f-494b-aa90-0bd032106c7b","Type":"ContainerStarted","Data":"120521b65276b0c232404200a1cc21e00880668850bb93c30a14b3eafe66fd03"} Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.176548 4931 generic.go:334] "Generic (PLEG): container finished" podID="c036f4c8-f544-4f0d-b1c6-51944420430c" containerID="a525778e298f6bc14bc2e734f045a374ff7899c5bd6f4b8d16546fe8bae5cf2d" exitCode=0 Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.176606 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" event={"ID":"c036f4c8-f544-4f0d-b1c6-51944420430c","Type":"ContainerDied","Data":"a525778e298f6bc14bc2e734f045a374ff7899c5bd6f4b8d16546fe8bae5cf2d"} Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.178995 4931 generic.go:334] "Generic (PLEG): container finished" podID="60e97735-4431-4143-ba6d-ec246a23722b" containerID="2157b2864740e777e186079bdc4e0060b4b7200710abbff10582b1caaf76290d" exitCode=0 Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.179075 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"60e97735-4431-4143-ba6d-ec246a23722b","Type":"ContainerDied","Data":"2157b2864740e777e186079bdc4e0060b4b7200710abbff10582b1caaf76290d"} Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.215752 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.329757 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hpqqz" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.364782 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-84p7l"] Nov 25 21:53:45 crc kubenswrapper[4931]: W1125 21:53:45.380003 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42040525_2c1c_46bf_94e6_b23e05b7b3ae.slice/crio-15e88f532f02d2838b7bea1de2e767778a07c2b3ffaf1213185dc945fccff0ff WatchSource:0}: Error finding container 15e88f532f02d2838b7bea1de2e767778a07c2b3ffaf1213185dc945fccff0ff: Status 404 returned error can't find the container with id 15e88f532f02d2838b7bea1de2e767778a07c2b3ffaf1213185dc945fccff0ff Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.404506 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.479443 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5gt24"] Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.519039 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.528789 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.531083 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60e97735-4431-4143-ba6d-ec246a23722b-kube-api-access\") pod \"60e97735-4431-4143-ba6d-ec246a23722b\" (UID: \"60e97735-4431-4143-ba6d-ec246a23722b\") " Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.531241 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60e97735-4431-4143-ba6d-ec246a23722b-kubelet-dir\") pod \"60e97735-4431-4143-ba6d-ec246a23722b\" (UID: \"60e97735-4431-4143-ba6d-ec246a23722b\") " Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.531465 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60e97735-4431-4143-ba6d-ec246a23722b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "60e97735-4431-4143-ba6d-ec246a23722b" (UID: "60e97735-4431-4143-ba6d-ec246a23722b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.545509 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60e97735-4431-4143-ba6d-ec246a23722b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "60e97735-4431-4143-ba6d-ec246a23722b" (UID: "60e97735-4431-4143-ba6d-ec246a23722b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.632913 4931 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60e97735-4431-4143-ba6d-ec246a23722b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.632944 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60e97735-4431-4143-ba6d-ec246a23722b-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.963910 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:45 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:45 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:45 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:45 crc kubenswrapper[4931]: I1125 21:53:45.963956 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.190098 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" event={"ID":"42040525-2c1c-46bf-94e6-b23e05b7b3ae","Type":"ContainerStarted","Data":"15e88f532f02d2838b7bea1de2e767778a07c2b3ffaf1213185dc945fccff0ff"} Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.194073 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"60e97735-4431-4143-ba6d-ec246a23722b","Type":"ContainerDied","Data":"638468860ccb77aa5d45c59744578aa6f0f1b3198ae7000fd826102c19c001e6"} Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.194113 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="638468860ccb77aa5d45c59744578aa6f0f1b3198ae7000fd826102c19c001e6" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.194086 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.196344 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gt24" event={"ID":"1e53eac6-a437-438d-b5ba-6747277367b8","Type":"ContainerStarted","Data":"7974584153aa4b947690e0ae49309f1ee155cdf13303f07ebf25858f4f707813"} Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.426809 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.430465 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.448133 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c036f4c8-f544-4f0d-b1c6-51944420430c-config-volume\") pod \"c036f4c8-f544-4f0d-b1c6-51944420430c\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.448231 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c036f4c8-f544-4f0d-b1c6-51944420430c-secret-volume\") pod \"c036f4c8-f544-4f0d-b1c6-51944420430c\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.448281 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdfxg\" (UniqueName: \"kubernetes.io/projected/c036f4c8-f544-4f0d-b1c6-51944420430c-kube-api-access-jdfxg\") pod \"c036f4c8-f544-4f0d-b1c6-51944420430c\" (UID: \"c036f4c8-f544-4f0d-b1c6-51944420430c\") " Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.448421 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.449144 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c036f4c8-f544-4f0d-b1c6-51944420430c-config-volume" (OuterVolumeSpecName: "config-volume") pod "c036f4c8-f544-4f0d-b1c6-51944420430c" (UID: "c036f4c8-f544-4f0d-b1c6-51944420430c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.452795 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c036f4c8-f544-4f0d-b1c6-51944420430c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c036f4c8-f544-4f0d-b1c6-51944420430c" (UID: "c036f4c8-f544-4f0d-b1c6-51944420430c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.454108 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/779e1587-ca2f-4330-91dd-7251714f63bc-metrics-certs\") pod \"network-metrics-daemon-7v4p9\" (UID: \"779e1587-ca2f-4330-91dd-7251714f63bc\") " pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.455380 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c036f4c8-f544-4f0d-b1c6-51944420430c-kube-api-access-jdfxg" (OuterVolumeSpecName: "kube-api-access-jdfxg") pod "c036f4c8-f544-4f0d-b1c6-51944420430c" (UID: "c036f4c8-f544-4f0d-b1c6-51944420430c"). InnerVolumeSpecName "kube-api-access-jdfxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.549823 4931 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c036f4c8-f544-4f0d-b1c6-51944420430c-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.549853 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdfxg\" (UniqueName: \"kubernetes.io/projected/c036f4c8-f544-4f0d-b1c6-51944420430c-kube-api-access-jdfxg\") on node \"crc\" DevicePath \"\"" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.549866 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c036f4c8-f544-4f0d-b1c6-51944420430c-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.575833 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 21:53:46 crc kubenswrapper[4931]: E1125 21:53:46.576046 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c036f4c8-f544-4f0d-b1c6-51944420430c" containerName="collect-profiles" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.576058 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c036f4c8-f544-4f0d-b1c6-51944420430c" containerName="collect-profiles" Nov 25 21:53:46 crc kubenswrapper[4931]: E1125 21:53:46.576077 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e97735-4431-4143-ba6d-ec246a23722b" containerName="pruner" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.576083 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e97735-4431-4143-ba6d-ec246a23722b" containerName="pruner" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.576174 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="c036f4c8-f544-4f0d-b1c6-51944420430c" containerName="collect-profiles" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.576186 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e97735-4431-4143-ba6d-ec246a23722b" containerName="pruner" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.576509 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.577258 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.578506 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.578547 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.651372 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.651415 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.660025 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7v4p9" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.752458 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.752522 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.752642 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.767021 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.896377 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.966694 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:46 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:46 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:46 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:46 crc kubenswrapper[4931]: I1125 21:53:46.967021 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.128011 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7v4p9"] Nov 25 21:53:47 crc kubenswrapper[4931]: W1125 21:53:47.133412 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod779e1587_ca2f_4330_91dd_7251714f63bc.slice/crio-8b7c2d608ac8530910ec3c3b652c47d493ae78b1228a1cf3a58a2908afaa6d9b WatchSource:0}: Error finding container 8b7c2d608ac8530910ec3c3b652c47d493ae78b1228a1cf3a58a2908afaa6d9b: Status 404 returned error can't find the container with id 8b7c2d608ac8530910ec3c3b652c47d493ae78b1228a1cf3a58a2908afaa6d9b Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.205098 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" event={"ID":"779e1587-ca2f-4330-91dd-7251714f63bc","Type":"ContainerStarted","Data":"8b7c2d608ac8530910ec3c3b652c47d493ae78b1228a1cf3a58a2908afaa6d9b"} Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.207157 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bgfgw" event={"ID":"6e742d61-fa3f-494b-aa90-0bd032106c7b","Type":"ContainerDied","Data":"030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50"} Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.207848 4931 generic.go:334] "Generic (PLEG): container finished" podID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerID="030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50" exitCode=0 Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.215017 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" event={"ID":"c036f4c8-f544-4f0d-b1c6-51944420430c","Type":"ContainerDied","Data":"ab001ca2bf3618ebc8f97756c6e5cfd925322969822205db12584758b31c7066"} Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.215058 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab001ca2bf3618ebc8f97756c6e5cfd925322969822205db12584758b31c7066" Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.215114 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401785-v5cgr" Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.231672 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-hxhhw" Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.359153 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 21:53:47 crc kubenswrapper[4931]: W1125 21:53:47.378570 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0f4ee33d_f7ee_4f5e_8178_644158f8a62e.slice/crio-8662d8a24c6155632ab9dde013e1a382a3d035e4ef7f764e080a1d68ce7d3313 WatchSource:0}: Error finding container 8662d8a24c6155632ab9dde013e1a382a3d035e4ef7f764e080a1d68ce7d3313: Status 404 returned error can't find the container with id 8662d8a24c6155632ab9dde013e1a382a3d035e4ef7f764e080a1d68ce7d3313 Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.963247 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:47 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:47 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:47 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:47 crc kubenswrapper[4931]: I1125 21:53:47.963843 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.221108 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" event={"ID":"42040525-2c1c-46bf-94e6-b23e05b7b3ae","Type":"ContainerStarted","Data":"8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f"} Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.221972 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.222993 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0f4ee33d-f7ee-4f5e-8178-644158f8a62e","Type":"ContainerStarted","Data":"8662d8a24c6155632ab9dde013e1a382a3d035e4ef7f764e080a1d68ce7d3313"} Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.225537 4931 generic.go:334] "Generic (PLEG): container finished" podID="1e53eac6-a437-438d-b5ba-6747277367b8" containerID="b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e" exitCode=0 Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.225669 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gt24" event={"ID":"1e53eac6-a437-438d-b5ba-6747277367b8","Type":"ContainerDied","Data":"b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e"} Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.245138 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" podStartSLOduration=144.245121164 podStartE2EDuration="2m24.245121164s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:48.241598158 +0000 UTC m=+166.598110277" watchObservedRunningTime="2025-11-25 21:53:48.245121164 +0000 UTC m=+166.601633283" Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.772261 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.777833 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-f8tsr" Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.962529 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:48 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:48 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:48 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:48 crc kubenswrapper[4931]: I1125 21:53:48.962596 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:49 crc kubenswrapper[4931]: I1125 21:53:49.243208 4931 generic.go:334] "Generic (PLEG): container finished" podID="0f4ee33d-f7ee-4f5e-8178-644158f8a62e" containerID="fcaa11f3c601e68f6e1d8a21a43d836c75b2ec7f896272f46e52b3ba806309a9" exitCode=0 Nov 25 21:53:49 crc kubenswrapper[4931]: I1125 21:53:49.243662 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0f4ee33d-f7ee-4f5e-8178-644158f8a62e","Type":"ContainerDied","Data":"fcaa11f3c601e68f6e1d8a21a43d836c75b2ec7f896272f46e52b3ba806309a9"} Nov 25 21:53:49 crc kubenswrapper[4931]: I1125 21:53:49.250741 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" event={"ID":"779e1587-ca2f-4330-91dd-7251714f63bc","Type":"ContainerStarted","Data":"384cfd540a3b49cc36ac123e8979d2487ed4bb99bac13f57fb87cbc0a862177b"} Nov 25 21:53:49 crc kubenswrapper[4931]: I1125 21:53:49.964417 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:49 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:49 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:49 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:49 crc kubenswrapper[4931]: I1125 21:53:49.965888 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:50 crc kubenswrapper[4931]: I1125 21:53:50.261522 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7v4p9" event={"ID":"779e1587-ca2f-4330-91dd-7251714f63bc","Type":"ContainerStarted","Data":"42095b9b26be44ff3a1b9586f93131a8fb308fe7a95d66a7c07907215b6ec5b8"} Nov 25 21:53:50 crc kubenswrapper[4931]: I1125 21:53:50.277406 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7v4p9" podStartSLOduration=146.277389285 podStartE2EDuration="2m26.277389285s" podCreationTimestamp="2025-11-25 21:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:53:50.276222553 +0000 UTC m=+168.632734682" watchObservedRunningTime="2025-11-25 21:53:50.277389285 +0000 UTC m=+168.633901404" Nov 25 21:53:50 crc kubenswrapper[4931]: I1125 21:53:50.962632 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:50 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:50 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:50 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:50 crc kubenswrapper[4931]: I1125 21:53:50.962974 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:51 crc kubenswrapper[4931]: I1125 21:53:51.962040 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:51 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:51 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:51 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:51 crc kubenswrapper[4931]: I1125 21:53:51.962086 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:52 crc kubenswrapper[4931]: I1125 21:53:52.588739 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:52 crc kubenswrapper[4931]: I1125 21:53:52.739481 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kubelet-dir\") pod \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\" (UID: \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\") " Nov 25 21:53:52 crc kubenswrapper[4931]: I1125 21:53:52.739530 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kube-api-access\") pod \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\" (UID: \"0f4ee33d-f7ee-4f5e-8178-644158f8a62e\") " Nov 25 21:53:52 crc kubenswrapper[4931]: I1125 21:53:52.739753 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0f4ee33d-f7ee-4f5e-8178-644158f8a62e" (UID: "0f4ee33d-f7ee-4f5e-8178-644158f8a62e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 21:53:52 crc kubenswrapper[4931]: I1125 21:53:52.751999 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0f4ee33d-f7ee-4f5e-8178-644158f8a62e" (UID: "0f4ee33d-f7ee-4f5e-8178-644158f8a62e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:53:52 crc kubenswrapper[4931]: I1125 21:53:52.841482 4931 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 21:53:52 crc kubenswrapper[4931]: I1125 21:53:52.841563 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f4ee33d-f7ee-4f5e-8178-644158f8a62e-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 21:53:52 crc kubenswrapper[4931]: I1125 21:53:52.962358 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:52 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:52 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:52 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:52 crc kubenswrapper[4931]: I1125 21:53:52.962418 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.281293 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0f4ee33d-f7ee-4f5e-8178-644158f8a62e","Type":"ContainerDied","Data":"8662d8a24c6155632ab9dde013e1a382a3d035e4ef7f764e080a1d68ce7d3313"} Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.281339 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8662d8a24c6155632ab9dde013e1a382a3d035e4ef7f764e080a1d68ce7d3313" Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.281415 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.749014 4931 patch_prober.go:28] interesting pod/console-f9d7485db-2hpxq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.749121 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2hpxq" podUID="69dfd797-b979-42a6-a4d5-d95e48bf7aa2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.779652 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.779724 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.779810 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.779865 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.962421 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:53 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:53 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:53 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:53 crc kubenswrapper[4931]: I1125 21:53:53.962483 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:54 crc kubenswrapper[4931]: I1125 21:53:54.962643 4931 patch_prober.go:28] interesting pod/router-default-5444994796-gdm7n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 21:53:54 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 25 21:53:54 crc kubenswrapper[4931]: [+]process-running ok Nov 25 21:53:54 crc kubenswrapper[4931]: healthz check failed Nov 25 21:53:54 crc kubenswrapper[4931]: I1125 21:53:54.962723 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gdm7n" podUID="ec936935-d216-4775-be35-f45ed5c5ceca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 21:53:55 crc kubenswrapper[4931]: I1125 21:53:55.964368 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:53:55 crc kubenswrapper[4931]: I1125 21:53:55.972744 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-gdm7n" Nov 25 21:54:03 crc kubenswrapper[4931]: I1125 21:54:03.778849 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:03 crc kubenswrapper[4931]: I1125 21:54:03.778915 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:03 crc kubenswrapper[4931]: I1125 21:54:03.779442 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:03 crc kubenswrapper[4931]: I1125 21:54:03.779496 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-9dxf2" Nov 25 21:54:03 crc kubenswrapper[4931]: I1125 21:54:03.779768 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:03 crc kubenswrapper[4931]: I1125 21:54:03.780656 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"8576fc2320aa7a788f2ffa24892f7f7ab4b1cf3fb07298f88e73a6a896b4f7e3"} pod="openshift-console/downloads-7954f5f757-9dxf2" containerMessage="Container download-server failed liveness probe, will be restarted" Nov 25 21:54:03 crc kubenswrapper[4931]: I1125 21:54:03.780744 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" containerID="cri-o://8576fc2320aa7a788f2ffa24892f7f7ab4b1cf3fb07298f88e73a6a896b4f7e3" gracePeriod=2 Nov 25 21:54:03 crc kubenswrapper[4931]: I1125 21:54:03.782102 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:03 crc kubenswrapper[4931]: I1125 21:54:03.782319 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:04 crc kubenswrapper[4931]: I1125 21:54:04.242452 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:54:04 crc kubenswrapper[4931]: I1125 21:54:04.248323 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-2hpxq" Nov 25 21:54:04 crc kubenswrapper[4931]: I1125 21:54:04.909721 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:54:05 crc kubenswrapper[4931]: I1125 21:54:05.362836 4931 generic.go:334] "Generic (PLEG): container finished" podID="7445c312-10f0-4296-ba35-2320e22382e8" containerID="8576fc2320aa7a788f2ffa24892f7f7ab4b1cf3fb07298f88e73a6a896b4f7e3" exitCode=0 Nov 25 21:54:05 crc kubenswrapper[4931]: I1125 21:54:05.362920 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9dxf2" event={"ID":"7445c312-10f0-4296-ba35-2320e22382e8","Type":"ContainerDied","Data":"8576fc2320aa7a788f2ffa24892f7f7ab4b1cf3fb07298f88e73a6a896b4f7e3"} Nov 25 21:54:06 crc kubenswrapper[4931]: E1125 21:54:06.169993 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 21:54:06 crc kubenswrapper[4931]: E1125 21:54:06.170187 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f6kks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qmg94_openshift-marketplace(ff6d71fd-060b-411c-a576-2651944470e3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 21:54:06 crc kubenswrapper[4931]: E1125 21:54:06.171609 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qmg94" podUID="ff6d71fd-060b-411c-a576-2651944470e3" Nov 25 21:54:08 crc kubenswrapper[4931]: I1125 21:54:08.843634 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 21:54:08 crc kubenswrapper[4931]: I1125 21:54:08.843699 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 21:54:10 crc kubenswrapper[4931]: E1125 21:54:10.564684 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 21:54:10 crc kubenswrapper[4931]: E1125 21:54:10.565613 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7bnt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-bnlql_openshift-marketplace(7d2bf749-d61d-4049-a314-dacdf107bdb5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 21:54:10 crc kubenswrapper[4931]: E1125 21:54:10.567058 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-bnlql" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" Nov 25 21:54:10 crc kubenswrapper[4931]: I1125 21:54:10.788804 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 21:54:13 crc kubenswrapper[4931]: I1125 21:54:13.778591 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:13 crc kubenswrapper[4931]: I1125 21:54:13.778653 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:15 crc kubenswrapper[4931]: I1125 21:54:15.323685 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bmxc6" Nov 25 21:54:18 crc kubenswrapper[4931]: E1125 21:54:18.260996 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 21:54:18 crc kubenswrapper[4931]: E1125 21:54:18.261226 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4n8ql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5gt24_openshift-marketplace(1e53eac6-a437-438d-b5ba-6747277367b8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 21:54:18 crc kubenswrapper[4931]: E1125 21:54:18.262527 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5gt24" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" Nov 25 21:54:23 crc kubenswrapper[4931]: I1125 21:54:23.782402 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:23 crc kubenswrapper[4931]: I1125 21:54:23.783421 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:33 crc kubenswrapper[4931]: I1125 21:54:33.779232 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:33 crc kubenswrapper[4931]: I1125 21:54:33.779974 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:38 crc kubenswrapper[4931]: I1125 21:54:38.844314 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 21:54:38 crc kubenswrapper[4931]: I1125 21:54:38.846135 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 21:54:38 crc kubenswrapper[4931]: I1125 21:54:38.846216 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:54:38 crc kubenswrapper[4931]: I1125 21:54:38.846824 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916"} pod="openshift-machine-config-operator/machine-config-daemon-djjhg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 21:54:38 crc kubenswrapper[4931]: I1125 21:54:38.846951 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" containerID="cri-o://1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916" gracePeriod=600 Nov 25 21:54:39 crc kubenswrapper[4931]: I1125 21:54:39.577222 4931 generic.go:334] "Generic (PLEG): container finished" podID="e22448de-5122-43be-853a-081347c13cf2" containerID="1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916" exitCode=0 Nov 25 21:54:39 crc kubenswrapper[4931]: I1125 21:54:39.577372 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerDied","Data":"1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916"} Nov 25 21:54:43 crc kubenswrapper[4931]: I1125 21:54:43.778823 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:43 crc kubenswrapper[4931]: I1125 21:54:43.779288 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:45 crc kubenswrapper[4931]: E1125 21:54:45.901372 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 21:54:45 crc kubenswrapper[4931]: E1125 21:54:45.901775 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xtffm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rj5p7_openshift-marketplace(e82c08a2-5336-42fa-b297-c432ddd981f0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 21:54:45 crc kubenswrapper[4931]: E1125 21:54:45.903185 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rj5p7" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" Nov 25 21:54:45 crc kubenswrapper[4931]: E1125 21:54:45.917415 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 21:54:45 crc kubenswrapper[4931]: E1125 21:54:45.918558 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v4mjz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pc84d_openshift-marketplace(b5d6d339-fbd3-461f-8862-340f1d44c2f8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 21:54:45 crc kubenswrapper[4931]: E1125 21:54:45.920680 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pc84d" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" Nov 25 21:54:45 crc kubenswrapper[4931]: E1125 21:54:45.922347 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 21:54:45 crc kubenswrapper[4931]: E1125 21:54:45.922504 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5wgzn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-bgfgw_openshift-marketplace(6e742d61-fa3f-494b-aa90-0bd032106c7b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 21:54:45 crc kubenswrapper[4931]: E1125 21:54:45.923702 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-bgfgw" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" Nov 25 21:54:47 crc kubenswrapper[4931]: E1125 21:54:47.784174 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-pc84d" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.688907 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnlql" event={"ID":"7d2bf749-d61d-4049-a314-dacdf107bdb5","Type":"ContainerStarted","Data":"b31ca6b91ce2ad18bc7ea8d1f5a5a2edc0d3ef2390365c9929a39067d920f579"} Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.690971 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gt24" event={"ID":"1e53eac6-a437-438d-b5ba-6747277367b8","Type":"ContainerStarted","Data":"c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6"} Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.693006 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glk9x" event={"ID":"abaedc86-626a-4a22-8f88-1738e338e27f","Type":"ContainerStarted","Data":"4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81"} Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.696630 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"8bda06aaa46440b32781ab98d8c14255214eac602bdadd7dbf0b48e5eb47636a"} Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.698281 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9dxf2" event={"ID":"7445c312-10f0-4296-ba35-2320e22382e8","Type":"ContainerStarted","Data":"97f8a394b86b8c72aa85878bda6e4e2987d0e30ffec58cac7b75606407658888"} Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.698900 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.698925 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9dxf2" Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.698939 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.699972 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmg94" event={"ID":"ff6d71fd-060b-411c-a576-2651944470e3","Type":"ContainerStarted","Data":"6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f"} Nov 25 21:54:48 crc kubenswrapper[4931]: I1125 21:54:48.700981 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fn56g" event={"ID":"16d28da7-dc24-4c86-914b-1bf11af8c9fd","Type":"ContainerStarted","Data":"de0a7e87bcbbf7f166c12e50dfcb27cd926edb0e1982a0805a8ef7f4edfcc091"} Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.709539 4931 generic.go:334] "Generic (PLEG): container finished" podID="ff6d71fd-060b-411c-a576-2651944470e3" containerID="6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f" exitCode=0 Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.709663 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmg94" event={"ID":"ff6d71fd-060b-411c-a576-2651944470e3","Type":"ContainerDied","Data":"6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f"} Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.712979 4931 generic.go:334] "Generic (PLEG): container finished" podID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerID="de0a7e87bcbbf7f166c12e50dfcb27cd926edb0e1982a0805a8ef7f4edfcc091" exitCode=0 Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.713175 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fn56g" event={"ID":"16d28da7-dc24-4c86-914b-1bf11af8c9fd","Type":"ContainerDied","Data":"de0a7e87bcbbf7f166c12e50dfcb27cd926edb0e1982a0805a8ef7f4edfcc091"} Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.716254 4931 generic.go:334] "Generic (PLEG): container finished" podID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerID="b31ca6b91ce2ad18bc7ea8d1f5a5a2edc0d3ef2390365c9929a39067d920f579" exitCode=0 Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.716355 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnlql" event={"ID":"7d2bf749-d61d-4049-a314-dacdf107bdb5","Type":"ContainerDied","Data":"b31ca6b91ce2ad18bc7ea8d1f5a5a2edc0d3ef2390365c9929a39067d920f579"} Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.720514 4931 generic.go:334] "Generic (PLEG): container finished" podID="1e53eac6-a437-438d-b5ba-6747277367b8" containerID="c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6" exitCode=0 Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.720604 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gt24" event={"ID":"1e53eac6-a437-438d-b5ba-6747277367b8","Type":"ContainerDied","Data":"c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6"} Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.725329 4931 generic.go:334] "Generic (PLEG): container finished" podID="abaedc86-626a-4a22-8f88-1738e338e27f" containerID="4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81" exitCode=0 Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.725503 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glk9x" event={"ID":"abaedc86-626a-4a22-8f88-1738e338e27f","Type":"ContainerDied","Data":"4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81"} Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.726678 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:49 crc kubenswrapper[4931]: I1125 21:54:49.726739 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:50 crc kubenswrapper[4931]: I1125 21:54:50.732097 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-9dxf2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Nov 25 21:54:50 crc kubenswrapper[4931]: I1125 21:54:50.732575 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9dxf2" podUID="7445c312-10f0-4296-ba35-2320e22382e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Nov 25 21:54:53 crc kubenswrapper[4931]: I1125 21:54:53.788545 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-9dxf2" Nov 25 21:55:11 crc kubenswrapper[4931]: I1125 21:55:11.870775 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fn56g" event={"ID":"16d28da7-dc24-4c86-914b-1bf11af8c9fd","Type":"ContainerStarted","Data":"cc739167636fdef87c3efc7e950a90504327851130a431723d0a2c77ac227a3b"} Nov 25 21:55:19 crc kubenswrapper[4931]: I1125 21:55:19.917560 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glk9x" event={"ID":"abaedc86-626a-4a22-8f88-1738e338e27f","Type":"ContainerStarted","Data":"3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e"} Nov 25 21:55:19 crc kubenswrapper[4931]: I1125 21:55:19.943439 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fn56g" podStartSLOduration=19.921732955 podStartE2EDuration="1m38.943425826s" podCreationTimestamp="2025-11-25 21:53:41 +0000 UTC" firstStartedPulling="2025-11-25 21:53:43.061303682 +0000 UTC m=+161.417815801" lastFinishedPulling="2025-11-25 21:55:02.082996523 +0000 UTC m=+240.439508672" observedRunningTime="2025-11-25 21:55:12.910831551 +0000 UTC m=+251.267343720" watchObservedRunningTime="2025-11-25 21:55:19.943425826 +0000 UTC m=+258.299937945" Nov 25 21:55:19 crc kubenswrapper[4931]: I1125 21:55:19.943559 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-glk9x" podStartSLOduration=17.467157891 podStartE2EDuration="1m38.943555849s" podCreationTimestamp="2025-11-25 21:53:41 +0000 UTC" firstStartedPulling="2025-11-25 21:53:43.016008295 +0000 UTC m=+161.372520414" lastFinishedPulling="2025-11-25 21:55:04.492406213 +0000 UTC m=+242.848918372" observedRunningTime="2025-11-25 21:55:19.941081771 +0000 UTC m=+258.297593930" watchObservedRunningTime="2025-11-25 21:55:19.943555849 +0000 UTC m=+258.300067968" Nov 25 21:55:21 crc kubenswrapper[4931]: I1125 21:55:21.823839 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:55:21 crc kubenswrapper[4931]: I1125 21:55:21.823924 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:55:22 crc kubenswrapper[4931]: I1125 21:55:22.038945 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:55:22 crc kubenswrapper[4931]: I1125 21:55:22.039415 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:55:22 crc kubenswrapper[4931]: I1125 21:55:22.299588 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:55:22 crc kubenswrapper[4931]: I1125 21:55:22.301146 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:55:22 crc kubenswrapper[4931]: I1125 21:55:22.383950 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:55:23 crc kubenswrapper[4931]: I1125 21:55:23.543706 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fn56g"] Nov 25 21:55:23 crc kubenswrapper[4931]: I1125 21:55:23.950163 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc84d" event={"ID":"b5d6d339-fbd3-461f-8862-340f1d44c2f8","Type":"ContainerStarted","Data":"47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0"} Nov 25 21:55:23 crc kubenswrapper[4931]: I1125 21:55:23.950298 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fn56g" podUID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerName="registry-server" containerID="cri-o://cc739167636fdef87c3efc7e950a90504327851130a431723d0a2c77ac227a3b" gracePeriod=2 Nov 25 21:55:24 crc kubenswrapper[4931]: I1125 21:55:24.960416 4931 generic.go:334] "Generic (PLEG): container finished" podID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerID="47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0" exitCode=0 Nov 25 21:55:24 crc kubenswrapper[4931]: I1125 21:55:24.960508 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc84d" event={"ID":"b5d6d339-fbd3-461f-8862-340f1d44c2f8","Type":"ContainerDied","Data":"47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0"} Nov 25 21:55:25 crc kubenswrapper[4931]: I1125 21:55:25.972400 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gt24" event={"ID":"1e53eac6-a437-438d-b5ba-6747277367b8","Type":"ContainerStarted","Data":"4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8"} Nov 25 21:55:26 crc kubenswrapper[4931]: I1125 21:55:26.981188 4931 generic.go:334] "Generic (PLEG): container finished" podID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerID="ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040" exitCode=0 Nov 25 21:55:26 crc kubenswrapper[4931]: I1125 21:55:26.981258 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rj5p7" event={"ID":"e82c08a2-5336-42fa-b297-c432ddd981f0","Type":"ContainerDied","Data":"ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040"} Nov 25 21:55:26 crc kubenswrapper[4931]: I1125 21:55:26.984185 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bgfgw" event={"ID":"6e742d61-fa3f-494b-aa90-0bd032106c7b","Type":"ContainerStarted","Data":"9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7"} Nov 25 21:55:26 crc kubenswrapper[4931]: I1125 21:55:26.987202 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmg94" event={"ID":"ff6d71fd-060b-411c-a576-2651944470e3","Type":"ContainerStarted","Data":"287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7"} Nov 25 21:55:26 crc kubenswrapper[4931]: I1125 21:55:26.989749 4931 generic.go:334] "Generic (PLEG): container finished" podID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerID="cc739167636fdef87c3efc7e950a90504327851130a431723d0a2c77ac227a3b" exitCode=0 Nov 25 21:55:26 crc kubenswrapper[4931]: I1125 21:55:26.989826 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fn56g" event={"ID":"16d28da7-dc24-4c86-914b-1bf11af8c9fd","Type":"ContainerDied","Data":"cc739167636fdef87c3efc7e950a90504327851130a431723d0a2c77ac227a3b"} Nov 25 21:55:26 crc kubenswrapper[4931]: I1125 21:55:26.992800 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnlql" event={"ID":"7d2bf749-d61d-4049-a314-dacdf107bdb5","Type":"ContainerStarted","Data":"cf127098744248e8fdf2cf181ef5c685eb8d13b37e7279111ad03dd2b9aba597"} Nov 25 21:55:28 crc kubenswrapper[4931]: I1125 21:55:28.003681 4931 generic.go:334] "Generic (PLEG): container finished" podID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerID="9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7" exitCode=0 Nov 25 21:55:28 crc kubenswrapper[4931]: I1125 21:55:28.003803 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bgfgw" event={"ID":"6e742d61-fa3f-494b-aa90-0bd032106c7b","Type":"ContainerDied","Data":"9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7"} Nov 25 21:55:28 crc kubenswrapper[4931]: I1125 21:55:28.037415 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bnlql" podStartSLOduration=21.522132855 podStartE2EDuration="1m47.037395086s" podCreationTimestamp="2025-11-25 21:53:41 +0000 UTC" firstStartedPulling="2025-11-25 21:53:43.064007616 +0000 UTC m=+161.420519725" lastFinishedPulling="2025-11-25 21:55:08.579269787 +0000 UTC m=+246.935781956" observedRunningTime="2025-11-25 21:55:28.035486624 +0000 UTC m=+266.391998793" watchObservedRunningTime="2025-11-25 21:55:28.037395086 +0000 UTC m=+266.393907205" Nov 25 21:55:28 crc kubenswrapper[4931]: I1125 21:55:28.061219 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qmg94" podStartSLOduration=19.655455459 podStartE2EDuration="1m47.061196648s" podCreationTimestamp="2025-11-25 21:53:41 +0000 UTC" firstStartedPulling="2025-11-25 21:53:43.02560525 +0000 UTC m=+161.382117369" lastFinishedPulling="2025-11-25 21:55:10.431346409 +0000 UTC m=+248.787858558" observedRunningTime="2025-11-25 21:55:28.056697185 +0000 UTC m=+266.413209304" watchObservedRunningTime="2025-11-25 21:55:28.061196648 +0000 UTC m=+266.417708767" Nov 25 21:55:28 crc kubenswrapper[4931]: I1125 21:55:28.083591 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5gt24" podStartSLOduration=22.745200182 podStartE2EDuration="1m44.083566581s" podCreationTimestamp="2025-11-25 21:53:44 +0000 UTC" firstStartedPulling="2025-11-25 21:53:48.228697502 +0000 UTC m=+166.585209621" lastFinishedPulling="2025-11-25 21:55:09.567063871 +0000 UTC m=+247.923576020" observedRunningTime="2025-11-25 21:55:28.081748371 +0000 UTC m=+266.438260500" watchObservedRunningTime="2025-11-25 21:55:28.083566581 +0000 UTC m=+266.440078720" Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.557475 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.695708 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-catalog-content\") pod \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.696223 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-utilities\") pod \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.696302 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phgnm\" (UniqueName: \"kubernetes.io/projected/16d28da7-dc24-4c86-914b-1bf11af8c9fd-kube-api-access-phgnm\") pod \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\" (UID: \"16d28da7-dc24-4c86-914b-1bf11af8c9fd\") " Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.697518 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-utilities" (OuterVolumeSpecName: "utilities") pod "16d28da7-dc24-4c86-914b-1bf11af8c9fd" (UID: "16d28da7-dc24-4c86-914b-1bf11af8c9fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.707125 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16d28da7-dc24-4c86-914b-1bf11af8c9fd-kube-api-access-phgnm" (OuterVolumeSpecName: "kube-api-access-phgnm") pod "16d28da7-dc24-4c86-914b-1bf11af8c9fd" (UID: "16d28da7-dc24-4c86-914b-1bf11af8c9fd"). InnerVolumeSpecName "kube-api-access-phgnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.740457 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16d28da7-dc24-4c86-914b-1bf11af8c9fd" (UID: "16d28da7-dc24-4c86-914b-1bf11af8c9fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.798250 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.798289 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phgnm\" (UniqueName: \"kubernetes.io/projected/16d28da7-dc24-4c86-914b-1bf11af8c9fd-kube-api-access-phgnm\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:29 crc kubenswrapper[4931]: I1125 21:55:29.798304 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16d28da7-dc24-4c86-914b-1bf11af8c9fd-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:30 crc kubenswrapper[4931]: I1125 21:55:30.020154 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fn56g" event={"ID":"16d28da7-dc24-4c86-914b-1bf11af8c9fd","Type":"ContainerDied","Data":"b85a88fd87699a702704b283f1f6addc357034da91c932b0cb589ce0b4f39f62"} Nov 25 21:55:30 crc kubenswrapper[4931]: I1125 21:55:30.020239 4931 scope.go:117] "RemoveContainer" containerID="cc739167636fdef87c3efc7e950a90504327851130a431723d0a2c77ac227a3b" Nov 25 21:55:30 crc kubenswrapper[4931]: I1125 21:55:30.020818 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fn56g" Nov 25 21:55:30 crc kubenswrapper[4931]: I1125 21:55:30.074511 4931 scope.go:117] "RemoveContainer" containerID="de0a7e87bcbbf7f166c12e50dfcb27cd926edb0e1982a0805a8ef7f4edfcc091" Nov 25 21:55:30 crc kubenswrapper[4931]: I1125 21:55:30.093419 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fn56g"] Nov 25 21:55:30 crc kubenswrapper[4931]: I1125 21:55:30.096304 4931 scope.go:117] "RemoveContainer" containerID="2a2fc74dc769b0a35c7f235ffabe2e0935121ddb624c5b9d7fc3fe4bd14bd7f0" Nov 25 21:55:30 crc kubenswrapper[4931]: I1125 21:55:30.097311 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fn56g"] Nov 25 21:55:30 crc kubenswrapper[4931]: I1125 21:55:30.434567 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" path="/var/lib/kubelet/pods/16d28da7-dc24-4c86-914b-1bf11af8c9fd/volumes" Nov 25 21:55:31 crc kubenswrapper[4931]: I1125 21:55:31.422542 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:55:31 crc kubenswrapper[4931]: I1125 21:55:31.422757 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:55:31 crc kubenswrapper[4931]: I1125 21:55:31.493995 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:55:32 crc kubenswrapper[4931]: I1125 21:55:32.074060 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:55:32 crc kubenswrapper[4931]: I1125 21:55:32.102400 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:55:32 crc kubenswrapper[4931]: I1125 21:55:32.254342 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:55:32 crc kubenswrapper[4931]: I1125 21:55:32.254414 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:55:32 crc kubenswrapper[4931]: I1125 21:55:32.311844 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:55:33 crc kubenswrapper[4931]: I1125 21:55:33.096026 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:55:35 crc kubenswrapper[4931]: I1125 21:55:35.217014 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:55:35 crc kubenswrapper[4931]: I1125 21:55:35.217340 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:55:35 crc kubenswrapper[4931]: I1125 21:55:35.252797 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bnlql"] Nov 25 21:55:35 crc kubenswrapper[4931]: I1125 21:55:35.253396 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bnlql" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerName="registry-server" containerID="cri-o://cf127098744248e8fdf2cf181ef5c685eb8d13b37e7279111ad03dd2b9aba597" gracePeriod=2 Nov 25 21:55:35 crc kubenswrapper[4931]: I1125 21:55:35.284928 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:55:36 crc kubenswrapper[4931]: I1125 21:55:36.107964 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:55:38 crc kubenswrapper[4931]: I1125 21:55:37.851652 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5gt24"] Nov 25 21:55:38 crc kubenswrapper[4931]: I1125 21:55:38.082145 4931 generic.go:334] "Generic (PLEG): container finished" podID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerID="cf127098744248e8fdf2cf181ef5c685eb8d13b37e7279111ad03dd2b9aba597" exitCode=0 Nov 25 21:55:38 crc kubenswrapper[4931]: I1125 21:55:38.082210 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnlql" event={"ID":"7d2bf749-d61d-4049-a314-dacdf107bdb5","Type":"ContainerDied","Data":"cf127098744248e8fdf2cf181ef5c685eb8d13b37e7279111ad03dd2b9aba597"} Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.092374 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnlql" event={"ID":"7d2bf749-d61d-4049-a314-dacdf107bdb5","Type":"ContainerDied","Data":"3ea90d2fe461cd291d86ef50f25fcbd935cd22b210d85baaab5105136fb72cd0"} Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.092455 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ea90d2fe461cd291d86ef50f25fcbd935cd22b210d85baaab5105136fb72cd0" Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.092504 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5gt24" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" containerName="registry-server" containerID="cri-o://4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8" gracePeriod=2 Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.138842 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.241569 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bnt7\" (UniqueName: \"kubernetes.io/projected/7d2bf749-d61d-4049-a314-dacdf107bdb5-kube-api-access-7bnt7\") pod \"7d2bf749-d61d-4049-a314-dacdf107bdb5\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.242021 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-utilities\") pod \"7d2bf749-d61d-4049-a314-dacdf107bdb5\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.242114 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-catalog-content\") pod \"7d2bf749-d61d-4049-a314-dacdf107bdb5\" (UID: \"7d2bf749-d61d-4049-a314-dacdf107bdb5\") " Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.246966 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d2bf749-d61d-4049-a314-dacdf107bdb5-kube-api-access-7bnt7" (OuterVolumeSpecName: "kube-api-access-7bnt7") pod "7d2bf749-d61d-4049-a314-dacdf107bdb5" (UID: "7d2bf749-d61d-4049-a314-dacdf107bdb5"). InnerVolumeSpecName "kube-api-access-7bnt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.260194 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-utilities" (OuterVolumeSpecName: "utilities") pod "7d2bf749-d61d-4049-a314-dacdf107bdb5" (UID: "7d2bf749-d61d-4049-a314-dacdf107bdb5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.296580 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d2bf749-d61d-4049-a314-dacdf107bdb5" (UID: "7d2bf749-d61d-4049-a314-dacdf107bdb5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.343092 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bnt7\" (UniqueName: \"kubernetes.io/projected/7d2bf749-d61d-4049-a314-dacdf107bdb5-kube-api-access-7bnt7\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.343125 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:39 crc kubenswrapper[4931]: I1125 21:55:39.343134 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d2bf749-d61d-4049-a314-dacdf107bdb5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.046218 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.100239 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc84d" event={"ID":"b5d6d339-fbd3-461f-8862-340f1d44c2f8","Type":"ContainerStarted","Data":"d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70"} Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.102794 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rj5p7" event={"ID":"e82c08a2-5336-42fa-b297-c432ddd981f0","Type":"ContainerStarted","Data":"12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795"} Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.105056 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bgfgw" event={"ID":"6e742d61-fa3f-494b-aa90-0bd032106c7b","Type":"ContainerStarted","Data":"10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67"} Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.112251 4931 generic.go:334] "Generic (PLEG): container finished" podID="1e53eac6-a437-438d-b5ba-6747277367b8" containerID="4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8" exitCode=0 Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.112349 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gt24" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.112383 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnlql" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.112420 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gt24" event={"ID":"1e53eac6-a437-438d-b5ba-6747277367b8","Type":"ContainerDied","Data":"4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8"} Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.112459 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gt24" event={"ID":"1e53eac6-a437-438d-b5ba-6747277367b8","Type":"ContainerDied","Data":"7974584153aa4b947690e0ae49309f1ee155cdf13303f07ebf25858f4f707813"} Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.112484 4931 scope.go:117] "RemoveContainer" containerID="4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.129755 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pc84d" podStartSLOduration=2.47441774 podStartE2EDuration="1m57.129732234s" podCreationTimestamp="2025-11-25 21:53:43 +0000 UTC" firstStartedPulling="2025-11-25 21:53:45.146466677 +0000 UTC m=+163.502978796" lastFinishedPulling="2025-11-25 21:55:39.801781171 +0000 UTC m=+278.158293290" observedRunningTime="2025-11-25 21:55:40.127049942 +0000 UTC m=+278.483562081" watchObservedRunningTime="2025-11-25 21:55:40.129732234 +0000 UTC m=+278.486244353" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.140212 4931 scope.go:117] "RemoveContainer" containerID="c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.152381 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rj5p7" podStartSLOduration=3.567084957 podStartE2EDuration="1m57.152359085s" podCreationTimestamp="2025-11-25 21:53:43 +0000 UTC" firstStartedPulling="2025-11-25 21:53:46.199178468 +0000 UTC m=+164.555690587" lastFinishedPulling="2025-11-25 21:55:39.784452586 +0000 UTC m=+278.140964715" observedRunningTime="2025-11-25 21:55:40.146623407 +0000 UTC m=+278.503135526" watchObservedRunningTime="2025-11-25 21:55:40.152359085 +0000 UTC m=+278.508871204" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.157122 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n8ql\" (UniqueName: \"kubernetes.io/projected/1e53eac6-a437-438d-b5ba-6747277367b8-kube-api-access-4n8ql\") pod \"1e53eac6-a437-438d-b5ba-6747277367b8\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.157213 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-catalog-content\") pod \"1e53eac6-a437-438d-b5ba-6747277367b8\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.157236 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-utilities\") pod \"1e53eac6-a437-438d-b5ba-6747277367b8\" (UID: \"1e53eac6-a437-438d-b5ba-6747277367b8\") " Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.163851 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-utilities" (OuterVolumeSpecName: "utilities") pod "1e53eac6-a437-438d-b5ba-6747277367b8" (UID: "1e53eac6-a437-438d-b5ba-6747277367b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.168670 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e53eac6-a437-438d-b5ba-6747277367b8-kube-api-access-4n8ql" (OuterVolumeSpecName: "kube-api-access-4n8ql") pod "1e53eac6-a437-438d-b5ba-6747277367b8" (UID: "1e53eac6-a437-438d-b5ba-6747277367b8"). InnerVolumeSpecName "kube-api-access-4n8ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.168891 4931 scope.go:117] "RemoveContainer" containerID="b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.171140 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bgfgw" podStartSLOduration=4.615398906 podStartE2EDuration="1m56.171120729s" podCreationTimestamp="2025-11-25 21:53:44 +0000 UTC" firstStartedPulling="2025-11-25 21:53:48.228532308 +0000 UTC m=+166.585044427" lastFinishedPulling="2025-11-25 21:55:39.784254131 +0000 UTC m=+278.140766250" observedRunningTime="2025-11-25 21:55:40.169395552 +0000 UTC m=+278.525907671" watchObservedRunningTime="2025-11-25 21:55:40.171120729 +0000 UTC m=+278.527632848" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.187345 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bnlql"] Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.193709 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bnlql"] Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.217037 4931 scope.go:117] "RemoveContainer" containerID="4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8" Nov 25 21:55:40 crc kubenswrapper[4931]: E1125 21:55:40.218067 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8\": container with ID starting with 4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8 not found: ID does not exist" containerID="4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.218111 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8"} err="failed to get container status \"4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8\": rpc error: code = NotFound desc = could not find container \"4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8\": container with ID starting with 4f4d9e30d99d5039c72da6a48f86400f007f56bcfdbe5b4335db6d5b3e3991a8 not found: ID does not exist" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.218139 4931 scope.go:117] "RemoveContainer" containerID="c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6" Nov 25 21:55:40 crc kubenswrapper[4931]: E1125 21:55:40.218416 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6\": container with ID starting with c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6 not found: ID does not exist" containerID="c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.218444 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6"} err="failed to get container status \"c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6\": rpc error: code = NotFound desc = could not find container \"c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6\": container with ID starting with c9f7218e266613ea44002ac23b443d6d91e8488e0c7dd68871aa0b7f76cd46f6 not found: ID does not exist" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.218461 4931 scope.go:117] "RemoveContainer" containerID="b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e" Nov 25 21:55:40 crc kubenswrapper[4931]: E1125 21:55:40.218767 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e\": container with ID starting with b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e not found: ID does not exist" containerID="b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.218797 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e"} err="failed to get container status \"b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e\": rpc error: code = NotFound desc = could not find container \"b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e\": container with ID starting with b2011622258dd9c30a7c321af1099408998c46b35ea569e65655703f3602029e not found: ID does not exist" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.257526 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e53eac6-a437-438d-b5ba-6747277367b8" (UID: "1e53eac6-a437-438d-b5ba-6747277367b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.258142 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.258167 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e53eac6-a437-438d-b5ba-6747277367b8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.258179 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n8ql\" (UniqueName: \"kubernetes.io/projected/1e53eac6-a437-438d-b5ba-6747277367b8-kube-api-access-4n8ql\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.430505 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" path="/var/lib/kubelet/pods/7d2bf749-d61d-4049-a314-dacdf107bdb5/volumes" Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.441055 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5gt24"] Nov 25 21:55:40 crc kubenswrapper[4931]: I1125 21:55:40.444720 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5gt24"] Nov 25 21:55:42 crc kubenswrapper[4931]: I1125 21:55:42.438132 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" path="/var/lib/kubelet/pods/1e53eac6-a437-438d-b5ba-6747277367b8/volumes" Nov 25 21:55:43 crc kubenswrapper[4931]: I1125 21:55:43.829908 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:55:43 crc kubenswrapper[4931]: I1125 21:55:43.830938 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:55:43 crc kubenswrapper[4931]: I1125 21:55:43.872054 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:55:44 crc kubenswrapper[4931]: I1125 21:55:44.169769 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:55:44 crc kubenswrapper[4931]: I1125 21:55:44.223995 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:55:44 crc kubenswrapper[4931]: I1125 21:55:44.224045 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:55:44 crc kubenswrapper[4931]: I1125 21:55:44.272210 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:55:44 crc kubenswrapper[4931]: I1125 21:55:44.800731 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:55:44 crc kubenswrapper[4931]: I1125 21:55:44.800783 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:55:45 crc kubenswrapper[4931]: I1125 21:55:45.185087 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:55:45 crc kubenswrapper[4931]: I1125 21:55:45.851326 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bgfgw" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerName="registry-server" probeResult="failure" output=< Nov 25 21:55:45 crc kubenswrapper[4931]: timeout: failed to connect service ":50051" within 1s Nov 25 21:55:45 crc kubenswrapper[4931]: > Nov 25 21:55:46 crc kubenswrapper[4931]: I1125 21:55:46.646830 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rj5p7"] Nov 25 21:55:47 crc kubenswrapper[4931]: I1125 21:55:47.145032 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rj5p7" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerName="registry-server" containerID="cri-o://12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795" gracePeriod=2 Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.044487 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.160830 4931 generic.go:334] "Generic (PLEG): container finished" podID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerID="12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795" exitCode=0 Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.160893 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rj5p7" event={"ID":"e82c08a2-5336-42fa-b297-c432ddd981f0","Type":"ContainerDied","Data":"12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795"} Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.160929 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rj5p7" event={"ID":"e82c08a2-5336-42fa-b297-c432ddd981f0","Type":"ContainerDied","Data":"e3c883a54228a15f29cf4cd262ca8def6f898297e9dfc4fae8ab90bc52c2cfab"} Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.160953 4931 scope.go:117] "RemoveContainer" containerID="12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.161031 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rj5p7" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.172923 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-utilities\") pod \"e82c08a2-5336-42fa-b297-c432ddd981f0\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.173049 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtffm\" (UniqueName: \"kubernetes.io/projected/e82c08a2-5336-42fa-b297-c432ddd981f0-kube-api-access-xtffm\") pod \"e82c08a2-5336-42fa-b297-c432ddd981f0\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.173104 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-catalog-content\") pod \"e82c08a2-5336-42fa-b297-c432ddd981f0\" (UID: \"e82c08a2-5336-42fa-b297-c432ddd981f0\") " Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.173792 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-utilities" (OuterVolumeSpecName: "utilities") pod "e82c08a2-5336-42fa-b297-c432ddd981f0" (UID: "e82c08a2-5336-42fa-b297-c432ddd981f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.177789 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e82c08a2-5336-42fa-b297-c432ddd981f0-kube-api-access-xtffm" (OuterVolumeSpecName: "kube-api-access-xtffm") pod "e82c08a2-5336-42fa-b297-c432ddd981f0" (UID: "e82c08a2-5336-42fa-b297-c432ddd981f0"). InnerVolumeSpecName "kube-api-access-xtffm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.180727 4931 scope.go:117] "RemoveContainer" containerID="ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.228602 4931 scope.go:117] "RemoveContainer" containerID="df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.231541 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e82c08a2-5336-42fa-b297-c432ddd981f0" (UID: "e82c08a2-5336-42fa-b297-c432ddd981f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.247724 4931 scope.go:117] "RemoveContainer" containerID="12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795" Nov 25 21:55:48 crc kubenswrapper[4931]: E1125 21:55:48.248303 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795\": container with ID starting with 12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795 not found: ID does not exist" containerID="12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.248363 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795"} err="failed to get container status \"12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795\": rpc error: code = NotFound desc = could not find container \"12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795\": container with ID starting with 12479fee38abb65a33e5faeb636c12c9b450e87616d4a06d201bbf86a350d795 not found: ID does not exist" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.248396 4931 scope.go:117] "RemoveContainer" containerID="ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040" Nov 25 21:55:48 crc kubenswrapper[4931]: E1125 21:55:48.248927 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040\": container with ID starting with ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040 not found: ID does not exist" containerID="ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.249000 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040"} err="failed to get container status \"ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040\": rpc error: code = NotFound desc = could not find container \"ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040\": container with ID starting with ff1c328584b0da474f2ebdf13be45607e06db56757660b766111b76e2f6bf040 not found: ID does not exist" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.249033 4931 scope.go:117] "RemoveContainer" containerID="df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30" Nov 25 21:55:48 crc kubenswrapper[4931]: E1125 21:55:48.249425 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30\": container with ID starting with df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30 not found: ID does not exist" containerID="df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.249459 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30"} err="failed to get container status \"df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30\": rpc error: code = NotFound desc = could not find container \"df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30\": container with ID starting with df525d531d290fc1e12d89312498fa94ed9fa5a4e7cafd41a14968758cdd1c30 not found: ID does not exist" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.283081 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.283139 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82c08a2-5336-42fa-b297-c432ddd981f0-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.283159 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtffm\" (UniqueName: \"kubernetes.io/projected/e82c08a2-5336-42fa-b297-c432ddd981f0-kube-api-access-xtffm\") on node \"crc\" DevicePath \"\"" Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.480527 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rj5p7"] Nov 25 21:55:48 crc kubenswrapper[4931]: I1125 21:55:48.483129 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rj5p7"] Nov 25 21:55:50 crc kubenswrapper[4931]: I1125 21:55:50.430635 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" path="/var/lib/kubelet/pods/e82c08a2-5336-42fa-b297-c432ddd981f0/volumes" Nov 25 21:55:54 crc kubenswrapper[4931]: I1125 21:55:54.840658 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:55:54 crc kubenswrapper[4931]: I1125 21:55:54.878088 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:56:02 crc kubenswrapper[4931]: I1125 21:56:02.166493 4931 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 25 21:56:25 crc kubenswrapper[4931]: I1125 21:56:25.931124 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p4p2g"] Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.776742 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qmg94"] Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.777195 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qmg94" podUID="ff6d71fd-060b-411c-a576-2651944470e3" containerName="registry-server" containerID="cri-o://287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7" gracePeriod=30 Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.785887 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-glk9x"] Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.786118 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-glk9x" podUID="abaedc86-626a-4a22-8f88-1738e338e27f" containerName="registry-server" containerID="cri-o://3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e" gracePeriod=30 Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.792322 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hsz8c"] Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.792564 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" podUID="c2a5d14a-4296-4064-8267-cf447ce6e41f" containerName="marketplace-operator" containerID="cri-o://524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f" gracePeriod=30 Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.819150 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc84d"] Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.819509 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pc84d" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerName="registry-server" containerID="cri-o://d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70" gracePeriod=30 Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842194 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hffz8"] Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842555 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerName="extract-utilities" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842616 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerName="extract-utilities" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842627 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerName="extract-content" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842633 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerName="extract-content" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842643 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842680 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842695 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842701 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842708 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerName="extract-content" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842714 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerName="extract-content" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842723 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerName="extract-utilities" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842775 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerName="extract-utilities" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842785 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842795 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842805 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerName="extract-content" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842835 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerName="extract-content" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842912 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4ee33d-f7ee-4f5e-8178-644158f8a62e" containerName="pruner" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842940 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4ee33d-f7ee-4f5e-8178-644158f8a62e" containerName="pruner" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.842956 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerName="extract-utilities" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.842995 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerName="extract-utilities" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.843010 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" containerName="extract-utilities" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843016 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" containerName="extract-utilities" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.843023 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" containerName="extract-content" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843030 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" containerName="extract-content" Nov 25 21:56:26 crc kubenswrapper[4931]: E1125 21:56:26.843069 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843078 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843236 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d2bf749-d61d-4049-a314-dacdf107bdb5" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843250 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e53eac6-a437-438d-b5ba-6747277367b8" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843259 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e82c08a2-5336-42fa-b297-c432ddd981f0" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843269 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f4ee33d-f7ee-4f5e-8178-644158f8a62e" containerName="pruner" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843282 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="16d28da7-dc24-4c86-914b-1bf11af8c9fd" containerName="registry-server" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843771 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.843949 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bgfgw"] Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.844215 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bgfgw" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerName="registry-server" containerID="cri-o://10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67" gracePeriod=30 Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.847280 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hffz8"] Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.966946 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpl82\" (UniqueName: \"kubernetes.io/projected/d0b65dd4-c1b3-4175-9951-3952c2fc4186-kube-api-access-tpl82\") pod \"marketplace-operator-79b997595-hffz8\" (UID: \"d0b65dd4-c1b3-4175-9951-3952c2fc4186\") " pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.967293 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d0b65dd4-c1b3-4175-9951-3952c2fc4186-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hffz8\" (UID: \"d0b65dd4-c1b3-4175-9951-3952c2fc4186\") " pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:26 crc kubenswrapper[4931]: I1125 21:56:26.967323 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0b65dd4-c1b3-4175-9951-3952c2fc4186-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hffz8\" (UID: \"d0b65dd4-c1b3-4175-9951-3952c2fc4186\") " pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.068723 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d0b65dd4-c1b3-4175-9951-3952c2fc4186-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hffz8\" (UID: \"d0b65dd4-c1b3-4175-9951-3952c2fc4186\") " pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.068788 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0b65dd4-c1b3-4175-9951-3952c2fc4186-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hffz8\" (UID: \"d0b65dd4-c1b3-4175-9951-3952c2fc4186\") " pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.068941 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpl82\" (UniqueName: \"kubernetes.io/projected/d0b65dd4-c1b3-4175-9951-3952c2fc4186-kube-api-access-tpl82\") pod \"marketplace-operator-79b997595-hffz8\" (UID: \"d0b65dd4-c1b3-4175-9951-3952c2fc4186\") " pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.071356 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0b65dd4-c1b3-4175-9951-3952c2fc4186-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hffz8\" (UID: \"d0b65dd4-c1b3-4175-9951-3952c2fc4186\") " pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.076394 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d0b65dd4-c1b3-4175-9951-3952c2fc4186-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hffz8\" (UID: \"d0b65dd4-c1b3-4175-9951-3952c2fc4186\") " pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.086527 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpl82\" (UniqueName: \"kubernetes.io/projected/d0b65dd4-c1b3-4175-9951-3952c2fc4186-kube-api-access-tpl82\") pod \"marketplace-operator-79b997595-hffz8\" (UID: \"d0b65dd4-c1b3-4175-9951-3952c2fc4186\") " pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.179536 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.244415 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.248535 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.256023 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.256450 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.272806 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-utilities\") pod \"abaedc86-626a-4a22-8f88-1738e338e27f\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.272864 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-operator-metrics\") pod \"c2a5d14a-4296-4064-8267-cf447ce6e41f\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.272948 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwv4c\" (UniqueName: \"kubernetes.io/projected/abaedc86-626a-4a22-8f88-1738e338e27f-kube-api-access-nwv4c\") pod \"abaedc86-626a-4a22-8f88-1738e338e27f\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.273108 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-catalog-content\") pod \"abaedc86-626a-4a22-8f88-1738e338e27f\" (UID: \"abaedc86-626a-4a22-8f88-1738e338e27f\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.273152 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-utilities\") pod \"ff6d71fd-060b-411c-a576-2651944470e3\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.273233 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-trusted-ca\") pod \"c2a5d14a-4296-4064-8267-cf447ce6e41f\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.273391 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-catalog-content\") pod \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.273417 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6kks\" (UniqueName: \"kubernetes.io/projected/ff6d71fd-060b-411c-a576-2651944470e3-kube-api-access-f6kks\") pod \"ff6d71fd-060b-411c-a576-2651944470e3\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.273448 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-catalog-content\") pod \"ff6d71fd-060b-411c-a576-2651944470e3\" (UID: \"ff6d71fd-060b-411c-a576-2651944470e3\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.273465 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4mjz\" (UniqueName: \"kubernetes.io/projected/b5d6d339-fbd3-461f-8862-340f1d44c2f8-kube-api-access-v4mjz\") pod \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.273574 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9hg4\" (UniqueName: \"kubernetes.io/projected/c2a5d14a-4296-4064-8267-cf447ce6e41f-kube-api-access-x9hg4\") pod \"c2a5d14a-4296-4064-8267-cf447ce6e41f\" (UID: \"c2a5d14a-4296-4064-8267-cf447ce6e41f\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.273646 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-utilities\") pod \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\" (UID: \"b5d6d339-fbd3-461f-8862-340f1d44c2f8\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.274719 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c2a5d14a-4296-4064-8267-cf447ce6e41f" (UID: "c2a5d14a-4296-4064-8267-cf447ce6e41f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.275395 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-utilities" (OuterVolumeSpecName: "utilities") pod "abaedc86-626a-4a22-8f88-1738e338e27f" (UID: "abaedc86-626a-4a22-8f88-1738e338e27f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.275506 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-utilities" (OuterVolumeSpecName: "utilities") pod "b5d6d339-fbd3-461f-8862-340f1d44c2f8" (UID: "b5d6d339-fbd3-461f-8862-340f1d44c2f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.279103 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-utilities" (OuterVolumeSpecName: "utilities") pod "ff6d71fd-060b-411c-a576-2651944470e3" (UID: "ff6d71fd-060b-411c-a576-2651944470e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.286224 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abaedc86-626a-4a22-8f88-1738e338e27f-kube-api-access-nwv4c" (OuterVolumeSpecName: "kube-api-access-nwv4c") pod "abaedc86-626a-4a22-8f88-1738e338e27f" (UID: "abaedc86-626a-4a22-8f88-1738e338e27f"). InnerVolumeSpecName "kube-api-access-nwv4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.287400 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.288263 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a5d14a-4296-4064-8267-cf447ce6e41f-kube-api-access-x9hg4" (OuterVolumeSpecName: "kube-api-access-x9hg4") pod "c2a5d14a-4296-4064-8267-cf447ce6e41f" (UID: "c2a5d14a-4296-4064-8267-cf447ce6e41f"). InnerVolumeSpecName "kube-api-access-x9hg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.288494 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c2a5d14a-4296-4064-8267-cf447ce6e41f" (UID: "c2a5d14a-4296-4064-8267-cf447ce6e41f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.294744 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff6d71fd-060b-411c-a576-2651944470e3-kube-api-access-f6kks" (OuterVolumeSpecName: "kube-api-access-f6kks") pod "ff6d71fd-060b-411c-a576-2651944470e3" (UID: "ff6d71fd-060b-411c-a576-2651944470e3"). InnerVolumeSpecName "kube-api-access-f6kks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.295394 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d6d339-fbd3-461f-8862-340f1d44c2f8-kube-api-access-v4mjz" (OuterVolumeSpecName: "kube-api-access-v4mjz") pod "b5d6d339-fbd3-461f-8862-340f1d44c2f8" (UID: "b5d6d339-fbd3-461f-8862-340f1d44c2f8"). InnerVolumeSpecName "kube-api-access-v4mjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.314920 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5d6d339-fbd3-461f-8862-340f1d44c2f8" (UID: "b5d6d339-fbd3-461f-8862-340f1d44c2f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.346197 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff6d71fd-060b-411c-a576-2651944470e3" (UID: "ff6d71fd-060b-411c-a576-2651944470e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.357382 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abaedc86-626a-4a22-8f88-1738e338e27f" (UID: "abaedc86-626a-4a22-8f88-1738e338e27f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374388 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-catalog-content\") pod \"6e742d61-fa3f-494b-aa90-0bd032106c7b\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374446 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-utilities\") pod \"6e742d61-fa3f-494b-aa90-0bd032106c7b\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374481 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wgzn\" (UniqueName: \"kubernetes.io/projected/6e742d61-fa3f-494b-aa90-0bd032106c7b-kube-api-access-5wgzn\") pod \"6e742d61-fa3f-494b-aa90-0bd032106c7b\" (UID: \"6e742d61-fa3f-494b-aa90-0bd032106c7b\") " Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374608 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374619 4931 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374631 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwv4c\" (UniqueName: \"kubernetes.io/projected/abaedc86-626a-4a22-8f88-1738e338e27f-kube-api-access-nwv4c\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374639 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaedc86-626a-4a22-8f88-1738e338e27f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374648 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374656 4931 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2a5d14a-4296-4064-8267-cf447ce6e41f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374664 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374673 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6kks\" (UniqueName: \"kubernetes.io/projected/ff6d71fd-060b-411c-a576-2651944470e3-kube-api-access-f6kks\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374680 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6d71fd-060b-411c-a576-2651944470e3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374688 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4mjz\" (UniqueName: \"kubernetes.io/projected/b5d6d339-fbd3-461f-8862-340f1d44c2f8-kube-api-access-v4mjz\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374696 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9hg4\" (UniqueName: \"kubernetes.io/projected/c2a5d14a-4296-4064-8267-cf447ce6e41f-kube-api-access-x9hg4\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.374704 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d6d339-fbd3-461f-8862-340f1d44c2f8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.375599 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-utilities" (OuterVolumeSpecName: "utilities") pod "6e742d61-fa3f-494b-aa90-0bd032106c7b" (UID: "6e742d61-fa3f-494b-aa90-0bd032106c7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.378065 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e742d61-fa3f-494b-aa90-0bd032106c7b-kube-api-access-5wgzn" (OuterVolumeSpecName: "kube-api-access-5wgzn") pod "6e742d61-fa3f-494b-aa90-0bd032106c7b" (UID: "6e742d61-fa3f-494b-aa90-0bd032106c7b"). InnerVolumeSpecName "kube-api-access-5wgzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.417084 4931 generic.go:334] "Generic (PLEG): container finished" podID="abaedc86-626a-4a22-8f88-1738e338e27f" containerID="3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e" exitCode=0 Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.417136 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glk9x" event={"ID":"abaedc86-626a-4a22-8f88-1738e338e27f","Type":"ContainerDied","Data":"3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.417161 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glk9x" event={"ID":"abaedc86-626a-4a22-8f88-1738e338e27f","Type":"ContainerDied","Data":"7cb61b740189b3a49b4e2ba7d6c0426b0077d76de421eeb4bb55258fdca9b3e0"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.417176 4931 scope.go:117] "RemoveContainer" containerID="3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.417265 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glk9x" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.424051 4931 generic.go:334] "Generic (PLEG): container finished" podID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerID="10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67" exitCode=0 Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.424120 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bgfgw" event={"ID":"6e742d61-fa3f-494b-aa90-0bd032106c7b","Type":"ContainerDied","Data":"10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.424160 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bgfgw" event={"ID":"6e742d61-fa3f-494b-aa90-0bd032106c7b","Type":"ContainerDied","Data":"120521b65276b0c232404200a1cc21e00880668850bb93c30a14b3eafe66fd03"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.424182 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bgfgw" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.432535 4931 generic.go:334] "Generic (PLEG): container finished" podID="ff6d71fd-060b-411c-a576-2651944470e3" containerID="287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7" exitCode=0 Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.432600 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmg94" event={"ID":"ff6d71fd-060b-411c-a576-2651944470e3","Type":"ContainerDied","Data":"287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.432608 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmg94" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.432628 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmg94" event={"ID":"ff6d71fd-060b-411c-a576-2651944470e3","Type":"ContainerDied","Data":"c30008e5c34bcdf9485f3cd53954e4b9e308486aed3a43d0bb9182f9d46f11e7"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.435694 4931 generic.go:334] "Generic (PLEG): container finished" podID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerID="d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70" exitCode=0 Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.435749 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pc84d" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.435834 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc84d" event={"ID":"b5d6d339-fbd3-461f-8862-340f1d44c2f8","Type":"ContainerDied","Data":"d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.435892 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc84d" event={"ID":"b5d6d339-fbd3-461f-8862-340f1d44c2f8","Type":"ContainerDied","Data":"04a46770fe8d445ca9c661ed848953ab3daee65e1fc3f985467aeb65b495c0a9"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.446242 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-glk9x"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.446599 4931 generic.go:334] "Generic (PLEG): container finished" podID="c2a5d14a-4296-4064-8267-cf447ce6e41f" containerID="524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f" exitCode=0 Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.446633 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" event={"ID":"c2a5d14a-4296-4064-8267-cf447ce6e41f","Type":"ContainerDied","Data":"524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.446658 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" event={"ID":"c2a5d14a-4296-4064-8267-cf447ce6e41f","Type":"ContainerDied","Data":"8d96b9e65d869ba2abccd66f54d696c92c730a7d4b08a98fda8cc2976eacaf3a"} Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.446703 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hsz8c" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.447011 4931 scope.go:117] "RemoveContainer" containerID="4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.465646 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-glk9x"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.469263 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qmg94"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.471784 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qmg94"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.476619 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wgzn\" (UniqueName: \"kubernetes.io/projected/6e742d61-fa3f-494b-aa90-0bd032106c7b-kube-api-access-5wgzn\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.476767 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.479359 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc84d"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.481072 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e742d61-fa3f-494b-aa90-0bd032106c7b" (UID: "6e742d61-fa3f-494b-aa90-0bd032106c7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.482080 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc84d"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.488329 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hsz8c"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.488705 4931 scope.go:117] "RemoveContainer" containerID="5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.490679 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hsz8c"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.511845 4931 scope.go:117] "RemoveContainer" containerID="3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.512718 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e\": container with ID starting with 3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e not found: ID does not exist" containerID="3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.512746 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e"} err="failed to get container status \"3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e\": rpc error: code = NotFound desc = could not find container \"3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e\": container with ID starting with 3ef0f79e9d5510645780c36be5fa42d1d28009b1f5897a51434fa77fee72000e not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.512768 4931 scope.go:117] "RemoveContainer" containerID="4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.513173 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81\": container with ID starting with 4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81 not found: ID does not exist" containerID="4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.513193 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81"} err="failed to get container status \"4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81\": rpc error: code = NotFound desc = could not find container \"4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81\": container with ID starting with 4a985bb30cec8fba2bca5d00126aa4cbfede4c011e41a7c6e2d9151aad8e0e81 not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.513205 4931 scope.go:117] "RemoveContainer" containerID="5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.513495 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f\": container with ID starting with 5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f not found: ID does not exist" containerID="5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.513523 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f"} err="failed to get container status \"5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f\": rpc error: code = NotFound desc = could not find container \"5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f\": container with ID starting with 5477be13053d40837582f540afe79b432e5cda73afa3cce3304fed67d1598f8f not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.513542 4931 scope.go:117] "RemoveContainer" containerID="10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.536258 4931 scope.go:117] "RemoveContainer" containerID="9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.565265 4931 scope.go:117] "RemoveContainer" containerID="030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.578610 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e742d61-fa3f-494b-aa90-0bd032106c7b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.582048 4931 scope.go:117] "RemoveContainer" containerID="10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.583432 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67\": container with ID starting with 10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67 not found: ID does not exist" containerID="10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.583474 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67"} err="failed to get container status \"10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67\": rpc error: code = NotFound desc = could not find container \"10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67\": container with ID starting with 10379d9baab242d87e2b1d1b156eca7befd6e0c9e74a3491ce19289dca53ca67 not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.583503 4931 scope.go:117] "RemoveContainer" containerID="9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.584178 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7\": container with ID starting with 9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7 not found: ID does not exist" containerID="9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.584232 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7"} err="failed to get container status \"9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7\": rpc error: code = NotFound desc = could not find container \"9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7\": container with ID starting with 9d1324a0c2c0f9ea5ded8722a6ebfe0e64ec4a48a8762c2616ec710e295236e7 not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.584269 4931 scope.go:117] "RemoveContainer" containerID="030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.584663 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50\": container with ID starting with 030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50 not found: ID does not exist" containerID="030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.584687 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50"} err="failed to get container status \"030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50\": rpc error: code = NotFound desc = could not find container \"030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50\": container with ID starting with 030cc70bc4c8031a77244fce402181ce20978c91ea61faed9269e108da1a5b50 not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.584705 4931 scope.go:117] "RemoveContainer" containerID="287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.601271 4931 scope.go:117] "RemoveContainer" containerID="6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.623302 4931 scope.go:117] "RemoveContainer" containerID="4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.637136 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hffz8"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.642581 4931 scope.go:117] "RemoveContainer" containerID="287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.643044 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7\": container with ID starting with 287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7 not found: ID does not exist" containerID="287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.643098 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7"} err="failed to get container status \"287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7\": rpc error: code = NotFound desc = could not find container \"287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7\": container with ID starting with 287cc96d4c628a60c4d57389ad581fa7b30196e8b17023c5b1ab9d47e81513b7 not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.643132 4931 scope.go:117] "RemoveContainer" containerID="6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.643709 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f\": container with ID starting with 6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f not found: ID does not exist" containerID="6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.643768 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f"} err="failed to get container status \"6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f\": rpc error: code = NotFound desc = could not find container \"6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f\": container with ID starting with 6babce485a02ca01564872ce710d4eb783f951413fef00e3ebc1ce24c4e0e69f not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.643790 4931 scope.go:117] "RemoveContainer" containerID="4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.644157 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96\": container with ID starting with 4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96 not found: ID does not exist" containerID="4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.644179 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96"} err="failed to get container status \"4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96\": rpc error: code = NotFound desc = could not find container \"4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96\": container with ID starting with 4ee1e1b98c7f35a6fa120373c1e0c60a39e16d654d3930b67e32ad7c957a8d96 not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.644193 4931 scope.go:117] "RemoveContainer" containerID="d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70" Nov 25 21:56:27 crc kubenswrapper[4931]: W1125 21:56:27.647496 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0b65dd4_c1b3_4175_9951_3952c2fc4186.slice/crio-ffd9e2d78dcc15e01864a9eed76dc27501664ec21cd2f312886dd464e91122e3 WatchSource:0}: Error finding container ffd9e2d78dcc15e01864a9eed76dc27501664ec21cd2f312886dd464e91122e3: Status 404 returned error can't find the container with id ffd9e2d78dcc15e01864a9eed76dc27501664ec21cd2f312886dd464e91122e3 Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.663315 4931 scope.go:117] "RemoveContainer" containerID="47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.677306 4931 scope.go:117] "RemoveContainer" containerID="eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.696019 4931 scope.go:117] "RemoveContainer" containerID="d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.696500 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70\": container with ID starting with d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70 not found: ID does not exist" containerID="d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.696542 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70"} err="failed to get container status \"d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70\": rpc error: code = NotFound desc = could not find container \"d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70\": container with ID starting with d11945790770e5709a708383b73b39739a08b7b7b7cd8a1d6fd9424341290f70 not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.696574 4931 scope.go:117] "RemoveContainer" containerID="47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.697059 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0\": container with ID starting with 47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0 not found: ID does not exist" containerID="47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.697082 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0"} err="failed to get container status \"47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0\": rpc error: code = NotFound desc = could not find container \"47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0\": container with ID starting with 47610930feadf078e7ba8f1af0f560d44ee6f8e8cc8fbbfed3be6e2ddd73dff0 not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.697097 4931 scope.go:117] "RemoveContainer" containerID="eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.697406 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c\": container with ID starting with eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c not found: ID does not exist" containerID="eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.697433 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c"} err="failed to get container status \"eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c\": rpc error: code = NotFound desc = could not find container \"eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c\": container with ID starting with eaf7551e513c9f5424258373558d3eef753e30fbc89c91863e19e60fcf462f4c not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.697448 4931 scope.go:117] "RemoveContainer" containerID="524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.720359 4931 scope.go:117] "RemoveContainer" containerID="524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.720745 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f\": container with ID starting with 524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f not found: ID does not exist" containerID="524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.720782 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f"} err="failed to get container status \"524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f\": rpc error: code = NotFound desc = could not find container \"524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f\": container with ID starting with 524f876003a11b3be66056e46be1893717881dc69f6eefc1f868aad4a455128f not found: ID does not exist" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.750089 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bgfgw"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.761094 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bgfgw"] Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.991954 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hhsmv"] Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992549 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaedc86-626a-4a22-8f88-1738e338e27f" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992570 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaedc86-626a-4a22-8f88-1738e338e27f" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992586 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerName="extract-utilities" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992597 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerName="extract-utilities" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992620 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6d71fd-060b-411c-a576-2651944470e3" containerName="extract-utilities" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992633 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6d71fd-060b-411c-a576-2651944470e3" containerName="extract-utilities" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992647 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaedc86-626a-4a22-8f88-1738e338e27f" containerName="extract-content" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992658 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaedc86-626a-4a22-8f88-1738e338e27f" containerName="extract-content" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992672 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6d71fd-060b-411c-a576-2651944470e3" containerName="extract-content" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992683 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6d71fd-060b-411c-a576-2651944470e3" containerName="extract-content" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992694 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992704 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992720 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2a5d14a-4296-4064-8267-cf447ce6e41f" containerName="marketplace-operator" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992730 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2a5d14a-4296-4064-8267-cf447ce6e41f" containerName="marketplace-operator" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992746 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaedc86-626a-4a22-8f88-1738e338e27f" containerName="extract-utilities" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992756 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaedc86-626a-4a22-8f88-1738e338e27f" containerName="extract-utilities" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992772 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992782 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992800 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerName="extract-utilities" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992810 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerName="extract-utilities" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992821 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerName="extract-content" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992831 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerName="extract-content" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992843 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6d71fd-060b-411c-a576-2651944470e3" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992853 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6d71fd-060b-411c-a576-2651944470e3" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: E1125 21:56:27.992891 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerName="extract-content" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.992903 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerName="extract-content" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.993045 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="abaedc86-626a-4a22-8f88-1738e338e27f" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.993064 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.993078 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2a5d14a-4296-4064-8267-cf447ce6e41f" containerName="marketplace-operator" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.993098 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.993111 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff6d71fd-060b-411c-a576-2651944470e3" containerName="registry-server" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.994119 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:27 crc kubenswrapper[4931]: I1125 21:56:27.996503 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.001619 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hhsmv"] Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.083815 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/372612f8-655b-4bbf-a36c-52b59ce3dda0-catalog-content\") pod \"community-operators-hhsmv\" (UID: \"372612f8-655b-4bbf-a36c-52b59ce3dda0\") " pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.083945 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/372612f8-655b-4bbf-a36c-52b59ce3dda0-utilities\") pod \"community-operators-hhsmv\" (UID: \"372612f8-655b-4bbf-a36c-52b59ce3dda0\") " pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.084006 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl5cb\" (UniqueName: \"kubernetes.io/projected/372612f8-655b-4bbf-a36c-52b59ce3dda0-kube-api-access-kl5cb\") pod \"community-operators-hhsmv\" (UID: \"372612f8-655b-4bbf-a36c-52b59ce3dda0\") " pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.186685 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/372612f8-655b-4bbf-a36c-52b59ce3dda0-utilities\") pod \"community-operators-hhsmv\" (UID: \"372612f8-655b-4bbf-a36c-52b59ce3dda0\") " pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.186749 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl5cb\" (UniqueName: \"kubernetes.io/projected/372612f8-655b-4bbf-a36c-52b59ce3dda0-kube-api-access-kl5cb\") pod \"community-operators-hhsmv\" (UID: \"372612f8-655b-4bbf-a36c-52b59ce3dda0\") " pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.186828 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/372612f8-655b-4bbf-a36c-52b59ce3dda0-catalog-content\") pod \"community-operators-hhsmv\" (UID: \"372612f8-655b-4bbf-a36c-52b59ce3dda0\") " pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.187129 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/372612f8-655b-4bbf-a36c-52b59ce3dda0-utilities\") pod \"community-operators-hhsmv\" (UID: \"372612f8-655b-4bbf-a36c-52b59ce3dda0\") " pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.187276 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/372612f8-655b-4bbf-a36c-52b59ce3dda0-catalog-content\") pod \"community-operators-hhsmv\" (UID: \"372612f8-655b-4bbf-a36c-52b59ce3dda0\") " pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.189110 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5g982"] Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.189992 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.191933 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.202024 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g982"] Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.207791 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl5cb\" (UniqueName: \"kubernetes.io/projected/372612f8-655b-4bbf-a36c-52b59ce3dda0-kube-api-access-kl5cb\") pod \"community-operators-hhsmv\" (UID: \"372612f8-655b-4bbf-a36c-52b59ce3dda0\") " pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.345733 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.388732 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a47b9365-6b25-4ce9-8f07-96afc03a8634-catalog-content\") pod \"redhat-marketplace-5g982\" (UID: \"a47b9365-6b25-4ce9-8f07-96afc03a8634\") " pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.388819 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a47b9365-6b25-4ce9-8f07-96afc03a8634-utilities\") pod \"redhat-marketplace-5g982\" (UID: \"a47b9365-6b25-4ce9-8f07-96afc03a8634\") " pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.388846 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcqhh\" (UniqueName: \"kubernetes.io/projected/a47b9365-6b25-4ce9-8f07-96afc03a8634-kube-api-access-dcqhh\") pod \"redhat-marketplace-5g982\" (UID: \"a47b9365-6b25-4ce9-8f07-96afc03a8634\") " pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.435925 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e742d61-fa3f-494b-aa90-0bd032106c7b" path="/var/lib/kubelet/pods/6e742d61-fa3f-494b-aa90-0bd032106c7b/volumes" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.436801 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abaedc86-626a-4a22-8f88-1738e338e27f" path="/var/lib/kubelet/pods/abaedc86-626a-4a22-8f88-1738e338e27f/volumes" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.437358 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5d6d339-fbd3-461f-8862-340f1d44c2f8" path="/var/lib/kubelet/pods/b5d6d339-fbd3-461f-8862-340f1d44c2f8/volumes" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.438323 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2a5d14a-4296-4064-8267-cf447ce6e41f" path="/var/lib/kubelet/pods/c2a5d14a-4296-4064-8267-cf447ce6e41f/volumes" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.438748 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff6d71fd-060b-411c-a576-2651944470e3" path="/var/lib/kubelet/pods/ff6d71fd-060b-411c-a576-2651944470e3/volumes" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.457048 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" event={"ID":"d0b65dd4-c1b3-4175-9951-3952c2fc4186","Type":"ContainerStarted","Data":"eb87536a2bcfa372ba61ea6efceb4d5debf9e912c0f80481fba7bc988fec041c"} Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.457095 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" event={"ID":"d0b65dd4-c1b3-4175-9951-3952c2fc4186","Type":"ContainerStarted","Data":"ffd9e2d78dcc15e01864a9eed76dc27501664ec21cd2f312886dd464e91122e3"} Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.457517 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.464637 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.478391 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" podStartSLOduration=2.478372621 podStartE2EDuration="2.478372621s" podCreationTimestamp="2025-11-25 21:56:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:56:28.475469122 +0000 UTC m=+326.831981251" watchObservedRunningTime="2025-11-25 21:56:28.478372621 +0000 UTC m=+326.834884730" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.491772 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a47b9365-6b25-4ce9-8f07-96afc03a8634-catalog-content\") pod \"redhat-marketplace-5g982\" (UID: \"a47b9365-6b25-4ce9-8f07-96afc03a8634\") " pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.491849 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a47b9365-6b25-4ce9-8f07-96afc03a8634-utilities\") pod \"redhat-marketplace-5g982\" (UID: \"a47b9365-6b25-4ce9-8f07-96afc03a8634\") " pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.491922 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcqhh\" (UniqueName: \"kubernetes.io/projected/a47b9365-6b25-4ce9-8f07-96afc03a8634-kube-api-access-dcqhh\") pod \"redhat-marketplace-5g982\" (UID: \"a47b9365-6b25-4ce9-8f07-96afc03a8634\") " pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.492717 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a47b9365-6b25-4ce9-8f07-96afc03a8634-catalog-content\") pod \"redhat-marketplace-5g982\" (UID: \"a47b9365-6b25-4ce9-8f07-96afc03a8634\") " pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.493234 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a47b9365-6b25-4ce9-8f07-96afc03a8634-utilities\") pod \"redhat-marketplace-5g982\" (UID: \"a47b9365-6b25-4ce9-8f07-96afc03a8634\") " pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.510589 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcqhh\" (UniqueName: \"kubernetes.io/projected/a47b9365-6b25-4ce9-8f07-96afc03a8634-kube-api-access-dcqhh\") pod \"redhat-marketplace-5g982\" (UID: \"a47b9365-6b25-4ce9-8f07-96afc03a8634\") " pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.759776 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hhsmv"] Nov 25 21:56:28 crc kubenswrapper[4931]: W1125 21:56:28.776027 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod372612f8_655b_4bbf_a36c_52b59ce3dda0.slice/crio-d1e29cb7813b64832d412ad546027cf70edd8ee3f57202c6841fa63fee7fcc18 WatchSource:0}: Error finding container d1e29cb7813b64832d412ad546027cf70edd8ee3f57202c6841fa63fee7fcc18: Status 404 returned error can't find the container with id d1e29cb7813b64832d412ad546027cf70edd8ee3f57202c6841fa63fee7fcc18 Nov 25 21:56:28 crc kubenswrapper[4931]: I1125 21:56:28.804549 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:29 crc kubenswrapper[4931]: I1125 21:56:29.010000 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g982"] Nov 25 21:56:29 crc kubenswrapper[4931]: I1125 21:56:29.469639 4931 generic.go:334] "Generic (PLEG): container finished" podID="372612f8-655b-4bbf-a36c-52b59ce3dda0" containerID="2ff590fff24db08fc1ffcaf5c473ba0eccff336c0679f9177c33dc6856cbd34a" exitCode=0 Nov 25 21:56:29 crc kubenswrapper[4931]: I1125 21:56:29.469734 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhsmv" event={"ID":"372612f8-655b-4bbf-a36c-52b59ce3dda0","Type":"ContainerDied","Data":"2ff590fff24db08fc1ffcaf5c473ba0eccff336c0679f9177c33dc6856cbd34a"} Nov 25 21:56:29 crc kubenswrapper[4931]: I1125 21:56:29.469792 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhsmv" event={"ID":"372612f8-655b-4bbf-a36c-52b59ce3dda0","Type":"ContainerStarted","Data":"d1e29cb7813b64832d412ad546027cf70edd8ee3f57202c6841fa63fee7fcc18"} Nov 25 21:56:29 crc kubenswrapper[4931]: I1125 21:56:29.472530 4931 generic.go:334] "Generic (PLEG): container finished" podID="a47b9365-6b25-4ce9-8f07-96afc03a8634" containerID="5ffa5396503dc2c04fbfebbcf3ab0b1433d3c9f03a30c958980f3f26c3153b9d" exitCode=0 Nov 25 21:56:29 crc kubenswrapper[4931]: I1125 21:56:29.472850 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g982" event={"ID":"a47b9365-6b25-4ce9-8f07-96afc03a8634","Type":"ContainerDied","Data":"5ffa5396503dc2c04fbfebbcf3ab0b1433d3c9f03a30c958980f3f26c3153b9d"} Nov 25 21:56:29 crc kubenswrapper[4931]: I1125 21:56:29.472971 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g982" event={"ID":"a47b9365-6b25-4ce9-8f07-96afc03a8634","Type":"ContainerStarted","Data":"a5f6668678c1120d8c555ae395390b4cf060fab9c0d0e6582ebe2856ae30647e"} Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.401975 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jhh6r"] Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.403307 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.405046 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.405367 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhh6r"] Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.480552 4931 generic.go:334] "Generic (PLEG): container finished" podID="a47b9365-6b25-4ce9-8f07-96afc03a8634" containerID="a2e69c8ccc598d753d21fc2437355abdcfb4ce3ce59fe2641db28630bed0009f" exitCode=0 Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.480739 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g982" event={"ID":"a47b9365-6b25-4ce9-8f07-96afc03a8634","Type":"ContainerDied","Data":"a2e69c8ccc598d753d21fc2437355abdcfb4ce3ce59fe2641db28630bed0009f"} Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.485364 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhsmv" event={"ID":"372612f8-655b-4bbf-a36c-52b59ce3dda0","Type":"ContainerStarted","Data":"306218b3ab3b7650ef5c4f8bb351d9e4ecc1758b7dd505941c1f3d3213430c85"} Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.516647 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddf049ad-4aee-4c0a-ab15-6c4ef560a755-catalog-content\") pod \"redhat-operators-jhh6r\" (UID: \"ddf049ad-4aee-4c0a-ab15-6c4ef560a755\") " pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.516745 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddf049ad-4aee-4c0a-ab15-6c4ef560a755-utilities\") pod \"redhat-operators-jhh6r\" (UID: \"ddf049ad-4aee-4c0a-ab15-6c4ef560a755\") " pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.516766 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dchpx\" (UniqueName: \"kubernetes.io/projected/ddf049ad-4aee-4c0a-ab15-6c4ef560a755-kube-api-access-dchpx\") pod \"redhat-operators-jhh6r\" (UID: \"ddf049ad-4aee-4c0a-ab15-6c4ef560a755\") " pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.592162 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fm7z9"] Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.593300 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.597766 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.611244 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fm7z9"] Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.620857 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dchpx\" (UniqueName: \"kubernetes.io/projected/ddf049ad-4aee-4c0a-ab15-6c4ef560a755-kube-api-access-dchpx\") pod \"redhat-operators-jhh6r\" (UID: \"ddf049ad-4aee-4c0a-ab15-6c4ef560a755\") " pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.621070 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddf049ad-4aee-4c0a-ab15-6c4ef560a755-catalog-content\") pod \"redhat-operators-jhh6r\" (UID: \"ddf049ad-4aee-4c0a-ab15-6c4ef560a755\") " pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.621354 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddf049ad-4aee-4c0a-ab15-6c4ef560a755-utilities\") pod \"redhat-operators-jhh6r\" (UID: \"ddf049ad-4aee-4c0a-ab15-6c4ef560a755\") " pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.621894 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddf049ad-4aee-4c0a-ab15-6c4ef560a755-utilities\") pod \"redhat-operators-jhh6r\" (UID: \"ddf049ad-4aee-4c0a-ab15-6c4ef560a755\") " pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.622863 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddf049ad-4aee-4c0a-ab15-6c4ef560a755-catalog-content\") pod \"redhat-operators-jhh6r\" (UID: \"ddf049ad-4aee-4c0a-ab15-6c4ef560a755\") " pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.644087 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dchpx\" (UniqueName: \"kubernetes.io/projected/ddf049ad-4aee-4c0a-ab15-6c4ef560a755-kube-api-access-dchpx\") pod \"redhat-operators-jhh6r\" (UID: \"ddf049ad-4aee-4c0a-ab15-6c4ef560a755\") " pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.722474 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbf2582-f183-485f-8947-7b8b5cd4dd75-utilities\") pod \"certified-operators-fm7z9\" (UID: \"acbf2582-f183-485f-8947-7b8b5cd4dd75\") " pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.722524 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbf2582-f183-485f-8947-7b8b5cd4dd75-catalog-content\") pod \"certified-operators-fm7z9\" (UID: \"acbf2582-f183-485f-8947-7b8b5cd4dd75\") " pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.722575 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hlxj\" (UniqueName: \"kubernetes.io/projected/acbf2582-f183-485f-8947-7b8b5cd4dd75-kube-api-access-4hlxj\") pod \"certified-operators-fm7z9\" (UID: \"acbf2582-f183-485f-8947-7b8b5cd4dd75\") " pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.746504 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.826690 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbf2582-f183-485f-8947-7b8b5cd4dd75-catalog-content\") pod \"certified-operators-fm7z9\" (UID: \"acbf2582-f183-485f-8947-7b8b5cd4dd75\") " pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.834245 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hlxj\" (UniqueName: \"kubernetes.io/projected/acbf2582-f183-485f-8947-7b8b5cd4dd75-kube-api-access-4hlxj\") pod \"certified-operators-fm7z9\" (UID: \"acbf2582-f183-485f-8947-7b8b5cd4dd75\") " pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.834403 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbf2582-f183-485f-8947-7b8b5cd4dd75-utilities\") pod \"certified-operators-fm7z9\" (UID: \"acbf2582-f183-485f-8947-7b8b5cd4dd75\") " pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.835020 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbf2582-f183-485f-8947-7b8b5cd4dd75-utilities\") pod \"certified-operators-fm7z9\" (UID: \"acbf2582-f183-485f-8947-7b8b5cd4dd75\") " pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.835768 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbf2582-f183-485f-8947-7b8b5cd4dd75-catalog-content\") pod \"certified-operators-fm7z9\" (UID: \"acbf2582-f183-485f-8947-7b8b5cd4dd75\") " pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.857554 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hlxj\" (UniqueName: \"kubernetes.io/projected/acbf2582-f183-485f-8947-7b8b5cd4dd75-kube-api-access-4hlxj\") pod \"certified-operators-fm7z9\" (UID: \"acbf2582-f183-485f-8947-7b8b5cd4dd75\") " pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.911706 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:30 crc kubenswrapper[4931]: I1125 21:56:30.942020 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhh6r"] Nov 25 21:56:30 crc kubenswrapper[4931]: W1125 21:56:30.953837 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddf049ad_4aee_4c0a_ab15_6c4ef560a755.slice/crio-335b707a180fde5c475c1a6ffad80635365acb357f864ffbaff6e5c99b53f272 WatchSource:0}: Error finding container 335b707a180fde5c475c1a6ffad80635365acb357f864ffbaff6e5c99b53f272: Status 404 returned error can't find the container with id 335b707a180fde5c475c1a6ffad80635365acb357f864ffbaff6e5c99b53f272 Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.094535 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fm7z9"] Nov 25 21:56:31 crc kubenswrapper[4931]: W1125 21:56:31.099186 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacbf2582_f183_485f_8947_7b8b5cd4dd75.slice/crio-a153b3bacabb0f3b9631f089e9ad96409c77c8405335bae1ce3b8a887c0e39fa WatchSource:0}: Error finding container a153b3bacabb0f3b9631f089e9ad96409c77c8405335bae1ce3b8a887c0e39fa: Status 404 returned error can't find the container with id a153b3bacabb0f3b9631f089e9ad96409c77c8405335bae1ce3b8a887c0e39fa Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.493428 4931 generic.go:334] "Generic (PLEG): container finished" podID="ddf049ad-4aee-4c0a-ab15-6c4ef560a755" containerID="8c5781129a60f5ab6faad783a7c0a80ed745f027366eaaf3e9c7cdeb6e3a370c" exitCode=0 Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.493492 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhh6r" event={"ID":"ddf049ad-4aee-4c0a-ab15-6c4ef560a755","Type":"ContainerDied","Data":"8c5781129a60f5ab6faad783a7c0a80ed745f027366eaaf3e9c7cdeb6e3a370c"} Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.493828 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhh6r" event={"ID":"ddf049ad-4aee-4c0a-ab15-6c4ef560a755","Type":"ContainerStarted","Data":"335b707a180fde5c475c1a6ffad80635365acb357f864ffbaff6e5c99b53f272"} Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.495955 4931 generic.go:334] "Generic (PLEG): container finished" podID="372612f8-655b-4bbf-a36c-52b59ce3dda0" containerID="306218b3ab3b7650ef5c4f8bb351d9e4ecc1758b7dd505941c1f3d3213430c85" exitCode=0 Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.496037 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhsmv" event={"ID":"372612f8-655b-4bbf-a36c-52b59ce3dda0","Type":"ContainerDied","Data":"306218b3ab3b7650ef5c4f8bb351d9e4ecc1758b7dd505941c1f3d3213430c85"} Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.498529 4931 generic.go:334] "Generic (PLEG): container finished" podID="acbf2582-f183-485f-8947-7b8b5cd4dd75" containerID="2e47d27b8d3f234a970cae233d18add6030c1c8812aa28a70aafc428bee56da5" exitCode=0 Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.499561 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm7z9" event={"ID":"acbf2582-f183-485f-8947-7b8b5cd4dd75","Type":"ContainerDied","Data":"2e47d27b8d3f234a970cae233d18add6030c1c8812aa28a70aafc428bee56da5"} Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.499599 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm7z9" event={"ID":"acbf2582-f183-485f-8947-7b8b5cd4dd75","Type":"ContainerStarted","Data":"a153b3bacabb0f3b9631f089e9ad96409c77c8405335bae1ce3b8a887c0e39fa"} Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.503516 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g982" event={"ID":"a47b9365-6b25-4ce9-8f07-96afc03a8634","Type":"ContainerStarted","Data":"cd766ccbcecb0afe22fc2c4488894c43c864392d57baf5ece659e43efcc53de2"} Nov 25 21:56:31 crc kubenswrapper[4931]: I1125 21:56:31.538245 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5g982" podStartSLOduration=2.1148954189999998 podStartE2EDuration="3.538229132s" podCreationTimestamp="2025-11-25 21:56:28 +0000 UTC" firstStartedPulling="2025-11-25 21:56:29.474861871 +0000 UTC m=+327.831373990" lastFinishedPulling="2025-11-25 21:56:30.898195554 +0000 UTC m=+329.254707703" observedRunningTime="2025-11-25 21:56:31.537399619 +0000 UTC m=+329.893911738" watchObservedRunningTime="2025-11-25 21:56:31.538229132 +0000 UTC m=+329.894741251" Nov 25 21:56:32 crc kubenswrapper[4931]: I1125 21:56:32.511809 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhsmv" event={"ID":"372612f8-655b-4bbf-a36c-52b59ce3dda0","Type":"ContainerStarted","Data":"a86f5787904282080736209955a02b537d96af683c3271cba50ff998545ab146"} Nov 25 21:56:32 crc kubenswrapper[4931]: I1125 21:56:32.514783 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhh6r" event={"ID":"ddf049ad-4aee-4c0a-ab15-6c4ef560a755","Type":"ContainerStarted","Data":"558a4a338f6010d47242137b48cd211596ef5814babe258d17507074e5ece8c7"} Nov 25 21:56:32 crc kubenswrapper[4931]: I1125 21:56:32.534920 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hhsmv" podStartSLOduration=3.109200004 podStartE2EDuration="5.534868046s" podCreationTimestamp="2025-11-25 21:56:27 +0000 UTC" firstStartedPulling="2025-11-25 21:56:29.473857494 +0000 UTC m=+327.830369653" lastFinishedPulling="2025-11-25 21:56:31.899525586 +0000 UTC m=+330.256037695" observedRunningTime="2025-11-25 21:56:32.531070562 +0000 UTC m=+330.887582701" watchObservedRunningTime="2025-11-25 21:56:32.534868046 +0000 UTC m=+330.891380205" Nov 25 21:56:33 crc kubenswrapper[4931]: I1125 21:56:33.527113 4931 generic.go:334] "Generic (PLEG): container finished" podID="ddf049ad-4aee-4c0a-ab15-6c4ef560a755" containerID="558a4a338f6010d47242137b48cd211596ef5814babe258d17507074e5ece8c7" exitCode=0 Nov 25 21:56:33 crc kubenswrapper[4931]: I1125 21:56:33.527169 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhh6r" event={"ID":"ddf049ad-4aee-4c0a-ab15-6c4ef560a755","Type":"ContainerDied","Data":"558a4a338f6010d47242137b48cd211596ef5814babe258d17507074e5ece8c7"} Nov 25 21:56:33 crc kubenswrapper[4931]: I1125 21:56:33.531280 4931 generic.go:334] "Generic (PLEG): container finished" podID="acbf2582-f183-485f-8947-7b8b5cd4dd75" containerID="f4e17e145831f910445e33dab73bbfe01a26bea241f62539b28e4cafc669cae6" exitCode=0 Nov 25 21:56:33 crc kubenswrapper[4931]: I1125 21:56:33.532326 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm7z9" event={"ID":"acbf2582-f183-485f-8947-7b8b5cd4dd75","Type":"ContainerDied","Data":"f4e17e145831f910445e33dab73bbfe01a26bea241f62539b28e4cafc669cae6"} Nov 25 21:56:34 crc kubenswrapper[4931]: I1125 21:56:34.539304 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhh6r" event={"ID":"ddf049ad-4aee-4c0a-ab15-6c4ef560a755","Type":"ContainerStarted","Data":"058fb6b6fb4836e9a1a92098703d5297004a73ef0d3bc722645cee50e4667c1c"} Nov 25 21:56:34 crc kubenswrapper[4931]: I1125 21:56:34.541288 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm7z9" event={"ID":"acbf2582-f183-485f-8947-7b8b5cd4dd75","Type":"ContainerStarted","Data":"db18be5fe5db6a7a6fbacb06c19c09700426c06cef86bd4ffe2a60c01c46a695"} Nov 25 21:56:34 crc kubenswrapper[4931]: I1125 21:56:34.563671 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jhh6r" podStartSLOduration=1.810329654 podStartE2EDuration="4.563652903s" podCreationTimestamp="2025-11-25 21:56:30 +0000 UTC" firstStartedPulling="2025-11-25 21:56:31.494810287 +0000 UTC m=+329.851322416" lastFinishedPulling="2025-11-25 21:56:34.248133546 +0000 UTC m=+332.604645665" observedRunningTime="2025-11-25 21:56:34.560044394 +0000 UTC m=+332.916556553" watchObservedRunningTime="2025-11-25 21:56:34.563652903 +0000 UTC m=+332.920165032" Nov 25 21:56:34 crc kubenswrapper[4931]: I1125 21:56:34.582108 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fm7z9" podStartSLOduration=2.098372621 podStartE2EDuration="4.582084246s" podCreationTimestamp="2025-11-25 21:56:30 +0000 UTC" firstStartedPulling="2025-11-25 21:56:31.500957285 +0000 UTC m=+329.857469404" lastFinishedPulling="2025-11-25 21:56:33.98466888 +0000 UTC m=+332.341181029" observedRunningTime="2025-11-25 21:56:34.576976556 +0000 UTC m=+332.933488685" watchObservedRunningTime="2025-11-25 21:56:34.582084246 +0000 UTC m=+332.938596405" Nov 25 21:56:38 crc kubenswrapper[4931]: I1125 21:56:38.347427 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:38 crc kubenswrapper[4931]: I1125 21:56:38.348935 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:38 crc kubenswrapper[4931]: I1125 21:56:38.398748 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:38 crc kubenswrapper[4931]: I1125 21:56:38.631223 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hhsmv" Nov 25 21:56:38 crc kubenswrapper[4931]: I1125 21:56:38.805383 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:38 crc kubenswrapper[4931]: I1125 21:56:38.805703 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:38 crc kubenswrapper[4931]: I1125 21:56:38.857774 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:39 crc kubenswrapper[4931]: I1125 21:56:39.622913 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5g982" Nov 25 21:56:40 crc kubenswrapper[4931]: I1125 21:56:40.746702 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:40 crc kubenswrapper[4931]: I1125 21:56:40.747202 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:40 crc kubenswrapper[4931]: I1125 21:56:40.813117 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:40 crc kubenswrapper[4931]: I1125 21:56:40.912052 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:40 crc kubenswrapper[4931]: I1125 21:56:40.912112 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:40 crc kubenswrapper[4931]: I1125 21:56:40.955665 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:41 crc kubenswrapper[4931]: I1125 21:56:41.618383 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jhh6r" Nov 25 21:56:41 crc kubenswrapper[4931]: I1125 21:56:41.625957 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fm7z9" Nov 25 21:56:50 crc kubenswrapper[4931]: I1125 21:56:50.960589 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" podUID="b364e7c6-bd4d-4d58-b348-c700933f4d47" containerName="oauth-openshift" containerID="cri-o://d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1" gracePeriod=15 Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.432358 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.462125 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-55c8c74798-dslfm"] Nov 25 21:56:51 crc kubenswrapper[4931]: E1125 21:56:51.462367 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b364e7c6-bd4d-4d58-b348-c700933f4d47" containerName="oauth-openshift" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.462381 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b364e7c6-bd4d-4d58-b348-c700933f4d47" containerName="oauth-openshift" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.462498 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b364e7c6-bd4d-4d58-b348-c700933f4d47" containerName="oauth-openshift" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.462935 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.478217 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-55c8c74798-dslfm"] Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530410 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530459 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-provider-selection\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530499 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530527 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-cliconfig\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530560 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-login\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530584 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-router-certs\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530617 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-service-ca\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530648 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-trusted-ca-bundle\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530668 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-session\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530684 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530843 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-ocp-branding-template\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530859 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-dir\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530891 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-error\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530920 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4ngn\" (UniqueName: \"kubernetes.io/projected/b364e7c6-bd4d-4d58-b348-c700933f4d47-kube-api-access-l4ngn\") pod \"b364e7c6-bd4d-4d58-b348-c700933f4d47\" (UID: \"b364e7c6-bd4d-4d58-b348-c700933f4d47\") " Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.530960 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531053 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531086 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531128 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-router-certs\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531147 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531212 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-audit-policies\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531261 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c7107d2-690c-4e5c-884a-1ab2b53dd362-audit-dir\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531284 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-service-ca\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531301 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llgqk\" (UniqueName: \"kubernetes.io/projected/8c7107d2-690c-4e5c-884a-1ab2b53dd362-kube-api-access-llgqk\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531333 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-template-login\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531369 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531395 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531411 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-template-error\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531434 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-session\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531462 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531506 4931 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531329 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531374 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.531816 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.533747 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.536336 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b364e7c6-bd4d-4d58-b348-c700933f4d47-kube-api-access-l4ngn" (OuterVolumeSpecName: "kube-api-access-l4ngn") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "kube-api-access-l4ngn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.537702 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.537911 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.538137 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.538297 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.538488 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.538609 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.538915 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.539157 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "b364e7c6-bd4d-4d58-b348-c700933f4d47" (UID: "b364e7c6-bd4d-4d58-b348-c700933f4d47"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.631971 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632047 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-audit-policies\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632076 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c7107d2-690c-4e5c-884a-1ab2b53dd362-audit-dir\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632105 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-service-ca\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632130 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llgqk\" (UniqueName: \"kubernetes.io/projected/8c7107d2-690c-4e5c-884a-1ab2b53dd362-kube-api-access-llgqk\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632160 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-template-login\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632189 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632216 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632313 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-template-error\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632341 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-session\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632394 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632428 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632454 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632482 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-router-certs\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632529 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632544 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632557 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632570 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632585 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632597 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632612 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632624 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632637 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4ngn\" (UniqueName: \"kubernetes.io/projected/b364e7c6-bd4d-4d58-b348-c700933f4d47-kube-api-access-l4ngn\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632650 4931 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632662 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632676 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.632688 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b364e7c6-bd4d-4d58-b348-c700933f4d47-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.633786 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c7107d2-690c-4e5c-884a-1ab2b53dd362-audit-dir\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.634925 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-audit-policies\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.635038 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.636075 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-service-ca\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.636766 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.637461 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.639323 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.640029 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.640563 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-session\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.642260 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-template-login\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.642492 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-router-certs\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.643376 4931 generic.go:334] "Generic (PLEG): container finished" podID="b364e7c6-bd4d-4d58-b348-c700933f4d47" containerID="d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1" exitCode=0 Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.643419 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" event={"ID":"b364e7c6-bd4d-4d58-b348-c700933f4d47","Type":"ContainerDied","Data":"d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1"} Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.643451 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" event={"ID":"b364e7c6-bd4d-4d58-b348-c700933f4d47","Type":"ContainerDied","Data":"29b7e4aa9bdaf8c6a4bf852ebcd52bcbc084001506fc1ca66cf23c97c3d2472c"} Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.643461 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-p4p2g" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.643723 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-user-template-error\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.643470 4931 scope.go:117] "RemoveContainer" containerID="d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.646914 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c7107d2-690c-4e5c-884a-1ab2b53dd362-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.668431 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llgqk\" (UniqueName: \"kubernetes.io/projected/8c7107d2-690c-4e5c-884a-1ab2b53dd362-kube-api-access-llgqk\") pod \"oauth-openshift-55c8c74798-dslfm\" (UID: \"8c7107d2-690c-4e5c-884a-1ab2b53dd362\") " pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.676455 4931 scope.go:117] "RemoveContainer" containerID="d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1" Nov 25 21:56:51 crc kubenswrapper[4931]: E1125 21:56:51.677183 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1\": container with ID starting with d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1 not found: ID does not exist" containerID="d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.677262 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1"} err="failed to get container status \"d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1\": rpc error: code = NotFound desc = could not find container \"d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1\": container with ID starting with d4487a367b93d6c62b3e257a2ebde59d3e7473c6873fb99da8b5f131a8ea83a1 not found: ID does not exist" Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.699231 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p4p2g"] Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.703091 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p4p2g"] Nov 25 21:56:51 crc kubenswrapper[4931]: I1125 21:56:51.786314 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:52 crc kubenswrapper[4931]: I1125 21:56:52.025194 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-55c8c74798-dslfm"] Nov 25 21:56:52 crc kubenswrapper[4931]: W1125 21:56:52.032223 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c7107d2_690c_4e5c_884a_1ab2b53dd362.slice/crio-56e4bdcc32e4bbbe93a1f64f626a445a9eeeb1370f635b9b7ed3f93a6dd86df4 WatchSource:0}: Error finding container 56e4bdcc32e4bbbe93a1f64f626a445a9eeeb1370f635b9b7ed3f93a6dd86df4: Status 404 returned error can't find the container with id 56e4bdcc32e4bbbe93a1f64f626a445a9eeeb1370f635b9b7ed3f93a6dd86df4 Nov 25 21:56:52 crc kubenswrapper[4931]: I1125 21:56:52.434544 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b364e7c6-bd4d-4d58-b348-c700933f4d47" path="/var/lib/kubelet/pods/b364e7c6-bd4d-4d58-b348-c700933f4d47/volumes" Nov 25 21:56:52 crc kubenswrapper[4931]: I1125 21:56:52.650797 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" event={"ID":"8c7107d2-690c-4e5c-884a-1ab2b53dd362","Type":"ContainerStarted","Data":"424af2def88cd1a337ffdfe250f7c838ebb4099796a4344956eba4765d49793f"} Nov 25 21:56:52 crc kubenswrapper[4931]: I1125 21:56:52.650862 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" event={"ID":"8c7107d2-690c-4e5c-884a-1ab2b53dd362","Type":"ContainerStarted","Data":"56e4bdcc32e4bbbe93a1f64f626a445a9eeeb1370f635b9b7ed3f93a6dd86df4"} Nov 25 21:56:52 crc kubenswrapper[4931]: I1125 21:56:52.651381 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:52 crc kubenswrapper[4931]: I1125 21:56:52.674416 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" Nov 25 21:56:52 crc kubenswrapper[4931]: I1125 21:56:52.699995 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-55c8c74798-dslfm" podStartSLOduration=27.699976008 podStartE2EDuration="27.699976008s" podCreationTimestamp="2025-11-25 21:56:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:56:52.678986376 +0000 UTC m=+351.035498495" watchObservedRunningTime="2025-11-25 21:56:52.699976008 +0000 UTC m=+351.056488127" Nov 25 21:57:08 crc kubenswrapper[4931]: I1125 21:57:08.844241 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 21:57:08 crc kubenswrapper[4931]: I1125 21:57:08.846129 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 21:57:38 crc kubenswrapper[4931]: I1125 21:57:38.845419 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 21:57:38 crc kubenswrapper[4931]: I1125 21:57:38.846143 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 21:58:08 crc kubenswrapper[4931]: I1125 21:58:08.844513 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 21:58:08 crc kubenswrapper[4931]: I1125 21:58:08.845222 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 21:58:08 crc kubenswrapper[4931]: I1125 21:58:08.845279 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 21:58:08 crc kubenswrapper[4931]: I1125 21:58:08.845947 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8bda06aaa46440b32781ab98d8c14255214eac602bdadd7dbf0b48e5eb47636a"} pod="openshift-machine-config-operator/machine-config-daemon-djjhg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 21:58:08 crc kubenswrapper[4931]: I1125 21:58:08.846006 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" containerID="cri-o://8bda06aaa46440b32781ab98d8c14255214eac602bdadd7dbf0b48e5eb47636a" gracePeriod=600 Nov 25 21:58:09 crc kubenswrapper[4931]: I1125 21:58:09.202096 4931 generic.go:334] "Generic (PLEG): container finished" podID="e22448de-5122-43be-853a-081347c13cf2" containerID="8bda06aaa46440b32781ab98d8c14255214eac602bdadd7dbf0b48e5eb47636a" exitCode=0 Nov 25 21:58:09 crc kubenswrapper[4931]: I1125 21:58:09.202186 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerDied","Data":"8bda06aaa46440b32781ab98d8c14255214eac602bdadd7dbf0b48e5eb47636a"} Nov 25 21:58:09 crc kubenswrapper[4931]: I1125 21:58:09.202522 4931 scope.go:117] "RemoveContainer" containerID="1c3d733e3cf41b9ba89228e7bf84447184970867cb1034a26b74216ef8eb7916" Nov 25 21:58:10 crc kubenswrapper[4931]: I1125 21:58:10.209833 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"e3f51442d6d51ca472d5b7468f6a60a5a87ed110feeee3ae340f88e811a0b5eb"} Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.816841 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8v66m"] Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.818289 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.830355 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8v66m"] Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.983672 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c1b5e8ab-c479-47c9-a634-422a6b23fc44-bound-sa-token\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.983766 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c1b5e8ab-c479-47c9-a634-422a6b23fc44-registry-certificates\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.983797 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.983833 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c1b5e8ab-c479-47c9-a634-422a6b23fc44-trusted-ca\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.983860 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c1b5e8ab-c479-47c9-a634-422a6b23fc44-registry-tls\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.983909 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c1b5e8ab-c479-47c9-a634-422a6b23fc44-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.983934 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdgq6\" (UniqueName: \"kubernetes.io/projected/c1b5e8ab-c479-47c9-a634-422a6b23fc44-kube-api-access-wdgq6\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:12 crc kubenswrapper[4931]: I1125 21:59:12.983954 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c1b5e8ab-c479-47c9-a634-422a6b23fc44-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.011783 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.085833 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c1b5e8ab-c479-47c9-a634-422a6b23fc44-registry-certificates\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.085951 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c1b5e8ab-c479-47c9-a634-422a6b23fc44-trusted-ca\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.086055 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c1b5e8ab-c479-47c9-a634-422a6b23fc44-registry-tls\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.086120 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c1b5e8ab-c479-47c9-a634-422a6b23fc44-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.086178 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdgq6\" (UniqueName: \"kubernetes.io/projected/c1b5e8ab-c479-47c9-a634-422a6b23fc44-kube-api-access-wdgq6\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.086228 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c1b5e8ab-c479-47c9-a634-422a6b23fc44-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.086362 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c1b5e8ab-c479-47c9-a634-422a6b23fc44-bound-sa-token\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.086771 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c1b5e8ab-c479-47c9-a634-422a6b23fc44-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.087911 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c1b5e8ab-c479-47c9-a634-422a6b23fc44-trusted-ca\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.088322 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c1b5e8ab-c479-47c9-a634-422a6b23fc44-registry-certificates\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.092627 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c1b5e8ab-c479-47c9-a634-422a6b23fc44-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.093191 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c1b5e8ab-c479-47c9-a634-422a6b23fc44-registry-tls\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.102977 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdgq6\" (UniqueName: \"kubernetes.io/projected/c1b5e8ab-c479-47c9-a634-422a6b23fc44-kube-api-access-wdgq6\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.103041 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c1b5e8ab-c479-47c9-a634-422a6b23fc44-bound-sa-token\") pod \"image-registry-66df7c8f76-8v66m\" (UID: \"c1b5e8ab-c479-47c9-a634-422a6b23fc44\") " pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.144446 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.355178 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8v66m"] Nov 25 21:59:13 crc kubenswrapper[4931]: W1125 21:59:13.359498 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1b5e8ab_c479_47c9_a634_422a6b23fc44.slice/crio-bf45175323e0ca39010d0853d36735c0130550fbb5bbb41c3b0da145af90790f WatchSource:0}: Error finding container bf45175323e0ca39010d0853d36735c0130550fbb5bbb41c3b0da145af90790f: Status 404 returned error can't find the container with id bf45175323e0ca39010d0853d36735c0130550fbb5bbb41c3b0da145af90790f Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.625757 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" event={"ID":"c1b5e8ab-c479-47c9-a634-422a6b23fc44","Type":"ContainerStarted","Data":"13f760a4f78b738a62c2c56b1b53a4bbf89ad065cba5b25d88db90e51d1dece0"} Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.625822 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" event={"ID":"c1b5e8ab-c479-47c9-a634-422a6b23fc44","Type":"ContainerStarted","Data":"bf45175323e0ca39010d0853d36735c0130550fbb5bbb41c3b0da145af90790f"} Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.625988 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:13 crc kubenswrapper[4931]: I1125 21:59:13.652162 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" podStartSLOduration=1.6521435819999999 podStartE2EDuration="1.652143582s" podCreationTimestamp="2025-11-25 21:59:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 21:59:13.65100045 +0000 UTC m=+492.007512589" watchObservedRunningTime="2025-11-25 21:59:13.652143582 +0000 UTC m=+492.008655711" Nov 25 21:59:33 crc kubenswrapper[4931]: I1125 21:59:33.158108 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8v66m" Nov 25 21:59:33 crc kubenswrapper[4931]: I1125 21:59:33.236914 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-84p7l"] Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.288284 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" podUID="42040525-2c1c-46bf-94e6-b23e05b7b3ae" containerName="registry" containerID="cri-o://8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f" gracePeriod=30 Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.740654 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.836443 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.836513 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42040525-2c1c-46bf-94e6-b23e05b7b3ae-installation-pull-secrets\") pod \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.836599 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-trusted-ca\") pod \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.836676 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-certificates\") pod \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.836716 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf6f8\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-kube-api-access-pf6f8\") pod \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.836750 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-tls\") pod \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.836809 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42040525-2c1c-46bf-94e6-b23e05b7b3ae-ca-trust-extracted\") pod \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.836842 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-bound-sa-token\") pod \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\" (UID: \"42040525-2c1c-46bf-94e6-b23e05b7b3ae\") " Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.838745 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "42040525-2c1c-46bf-94e6-b23e05b7b3ae" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.842451 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "42040525-2c1c-46bf-94e6-b23e05b7b3ae" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.843861 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "42040525-2c1c-46bf-94e6-b23e05b7b3ae" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.855572 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42040525-2c1c-46bf-94e6-b23e05b7b3ae-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "42040525-2c1c-46bf-94e6-b23e05b7b3ae" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.856371 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "42040525-2c1c-46bf-94e6-b23e05b7b3ae" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.857635 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-kube-api-access-pf6f8" (OuterVolumeSpecName: "kube-api-access-pf6f8") pod "42040525-2c1c-46bf-94e6-b23e05b7b3ae" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae"). InnerVolumeSpecName "kube-api-access-pf6f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.861432 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "42040525-2c1c-46bf-94e6-b23e05b7b3ae" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.867266 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42040525-2c1c-46bf-94e6-b23e05b7b3ae-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "42040525-2c1c-46bf-94e6-b23e05b7b3ae" (UID: "42040525-2c1c-46bf-94e6-b23e05b7b3ae"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.938138 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf6f8\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-kube-api-access-pf6f8\") on node \"crc\" DevicePath \"\"" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.938189 4931 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.938201 4931 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42040525-2c1c-46bf-94e6-b23e05b7b3ae-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.938212 4931 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42040525-2c1c-46bf-94e6-b23e05b7b3ae-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.938225 4931 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42040525-2c1c-46bf-94e6-b23e05b7b3ae-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.938236 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.938250 4931 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42040525-2c1c-46bf-94e6-b23e05b7b3ae-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.985390 4931 generic.go:334] "Generic (PLEG): container finished" podID="42040525-2c1c-46bf-94e6-b23e05b7b3ae" containerID="8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f" exitCode=0 Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.985456 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" event={"ID":"42040525-2c1c-46bf-94e6-b23e05b7b3ae","Type":"ContainerDied","Data":"8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f"} Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.985500 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" event={"ID":"42040525-2c1c-46bf-94e6-b23e05b7b3ae","Type":"ContainerDied","Data":"15e88f532f02d2838b7bea1de2e767778a07c2b3ffaf1213185dc945fccff0ff"} Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.985530 4931 scope.go:117] "RemoveContainer" containerID="8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f" Nov 25 21:59:58 crc kubenswrapper[4931]: I1125 21:59:58.985687 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-84p7l" Nov 25 21:59:59 crc kubenswrapper[4931]: I1125 21:59:59.015099 4931 scope.go:117] "RemoveContainer" containerID="8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f" Nov 25 21:59:59 crc kubenswrapper[4931]: E1125 21:59:59.016089 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f\": container with ID starting with 8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f not found: ID does not exist" containerID="8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f" Nov 25 21:59:59 crc kubenswrapper[4931]: I1125 21:59:59.016132 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f"} err="failed to get container status \"8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f\": rpc error: code = NotFound desc = could not find container \"8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f\": container with ID starting with 8423f5b817478eac41e8aafdf54cb1cf666faab64a6d2ad95aed8798e24d015f not found: ID does not exist" Nov 25 21:59:59 crc kubenswrapper[4931]: I1125 21:59:59.034721 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-84p7l"] Nov 25 21:59:59 crc kubenswrapper[4931]: I1125 21:59:59.041688 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-84p7l"] Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.140827 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv"] Nov 25 22:00:00 crc kubenswrapper[4931]: E1125 22:00:00.141471 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42040525-2c1c-46bf-94e6-b23e05b7b3ae" containerName="registry" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.141491 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="42040525-2c1c-46bf-94e6-b23e05b7b3ae" containerName="registry" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.141632 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="42040525-2c1c-46bf-94e6-b23e05b7b3ae" containerName="registry" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.142206 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.146519 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.147679 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.152944 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-config-volume\") pod \"collect-profiles-29401800-jd8tv\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.153009 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv"] Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.153022 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6ffx\" (UniqueName: \"kubernetes.io/projected/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-kube-api-access-w6ffx\") pod \"collect-profiles-29401800-jd8tv\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.153097 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-secret-volume\") pod \"collect-profiles-29401800-jd8tv\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.254403 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-secret-volume\") pod \"collect-profiles-29401800-jd8tv\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.254497 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-config-volume\") pod \"collect-profiles-29401800-jd8tv\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.254523 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6ffx\" (UniqueName: \"kubernetes.io/projected/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-kube-api-access-w6ffx\") pod \"collect-profiles-29401800-jd8tv\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.255408 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-config-volume\") pod \"collect-profiles-29401800-jd8tv\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.259159 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-secret-volume\") pod \"collect-profiles-29401800-jd8tv\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.268749 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6ffx\" (UniqueName: \"kubernetes.io/projected/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-kube-api-access-w6ffx\") pod \"collect-profiles-29401800-jd8tv\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.433234 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42040525-2c1c-46bf-94e6-b23e05b7b3ae" path="/var/lib/kubelet/pods/42040525-2c1c-46bf-94e6-b23e05b7b3ae/volumes" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.472917 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:00 crc kubenswrapper[4931]: I1125 22:00:00.663946 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv"] Nov 25 22:00:01 crc kubenswrapper[4931]: I1125 22:00:01.001679 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" event={"ID":"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7","Type":"ContainerStarted","Data":"db6440ec8a44f3b6010f5bfb09ed9e835183741276df2eb68d01d56363dfa6a2"} Nov 25 22:00:01 crc kubenswrapper[4931]: I1125 22:00:01.001756 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" event={"ID":"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7","Type":"ContainerStarted","Data":"abcedb9029351660357fe2de4bad7c07d5d60ea0ef16254ae503426f4c0bcb8e"} Nov 25 22:00:01 crc kubenswrapper[4931]: I1125 22:00:01.014518 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" podStartSLOduration=1.014501763 podStartE2EDuration="1.014501763s" podCreationTimestamp="2025-11-25 22:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:00:01.012660273 +0000 UTC m=+539.369172392" watchObservedRunningTime="2025-11-25 22:00:01.014501763 +0000 UTC m=+539.371013882" Nov 25 22:00:02 crc kubenswrapper[4931]: I1125 22:00:02.009299 4931 generic.go:334] "Generic (PLEG): container finished" podID="2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7" containerID="db6440ec8a44f3b6010f5bfb09ed9e835183741276df2eb68d01d56363dfa6a2" exitCode=0 Nov 25 22:00:02 crc kubenswrapper[4931]: I1125 22:00:02.009479 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" event={"ID":"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7","Type":"ContainerDied","Data":"db6440ec8a44f3b6010f5bfb09ed9e835183741276df2eb68d01d56363dfa6a2"} Nov 25 22:00:02 crc kubenswrapper[4931]: I1125 22:00:02.598804 4931 scope.go:117] "RemoveContainer" containerID="fefd4f318f057ab6277310db0ac2b7c615321f175488b48fc7bec511d778a3bf" Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.307600 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.398166 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-secret-volume\") pod \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.399095 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6ffx\" (UniqueName: \"kubernetes.io/projected/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-kube-api-access-w6ffx\") pod \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.399590 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-config-volume\") pod \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\" (UID: \"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7\") " Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.400162 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-config-volume" (OuterVolumeSpecName: "config-volume") pod "2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7" (UID: "2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.400768 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.404837 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-kube-api-access-w6ffx" (OuterVolumeSpecName: "kube-api-access-w6ffx") pod "2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7" (UID: "2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7"). InnerVolumeSpecName "kube-api-access-w6ffx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.404965 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7" (UID: "2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.502252 4931 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 22:00:03 crc kubenswrapper[4931]: I1125 22:00:03.502340 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6ffx\" (UniqueName: \"kubernetes.io/projected/2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7-kube-api-access-w6ffx\") on node \"crc\" DevicePath \"\"" Nov 25 22:00:04 crc kubenswrapper[4931]: I1125 22:00:04.026721 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" event={"ID":"2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7","Type":"ContainerDied","Data":"abcedb9029351660357fe2de4bad7c07d5d60ea0ef16254ae503426f4c0bcb8e"} Nov 25 22:00:04 crc kubenswrapper[4931]: I1125 22:00:04.026769 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abcedb9029351660357fe2de4bad7c07d5d60ea0ef16254ae503426f4c0bcb8e" Nov 25 22:00:04 crc kubenswrapper[4931]: I1125 22:00:04.026801 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401800-jd8tv" Nov 25 22:00:38 crc kubenswrapper[4931]: I1125 22:00:38.845347 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:00:38 crc kubenswrapper[4931]: I1125 22:00:38.846361 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:01:02 crc kubenswrapper[4931]: I1125 22:01:02.639935 4931 scope.go:117] "RemoveContainer" containerID="b31ca6b91ce2ad18bc7ea8d1f5a5a2edc0d3ef2390365c9929a39067d920f579" Nov 25 22:01:08 crc kubenswrapper[4931]: I1125 22:01:08.843508 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:01:08 crc kubenswrapper[4931]: I1125 22:01:08.844305 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:01:38 crc kubenswrapper[4931]: I1125 22:01:38.849260 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:01:38 crc kubenswrapper[4931]: I1125 22:01:38.849930 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:01:38 crc kubenswrapper[4931]: I1125 22:01:38.849995 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 22:01:38 crc kubenswrapper[4931]: I1125 22:01:38.850842 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e3f51442d6d51ca472d5b7468f6a60a5a87ed110feeee3ae340f88e811a0b5eb"} pod="openshift-machine-config-operator/machine-config-daemon-djjhg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 22:01:38 crc kubenswrapper[4931]: I1125 22:01:38.850958 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" containerID="cri-o://e3f51442d6d51ca472d5b7468f6a60a5a87ed110feeee3ae340f88e811a0b5eb" gracePeriod=600 Nov 25 22:01:39 crc kubenswrapper[4931]: I1125 22:01:39.672794 4931 generic.go:334] "Generic (PLEG): container finished" podID="e22448de-5122-43be-853a-081347c13cf2" containerID="e3f51442d6d51ca472d5b7468f6a60a5a87ed110feeee3ae340f88e811a0b5eb" exitCode=0 Nov 25 22:01:39 crc kubenswrapper[4931]: I1125 22:01:39.672936 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerDied","Data":"e3f51442d6d51ca472d5b7468f6a60a5a87ed110feeee3ae340f88e811a0b5eb"} Nov 25 22:01:39 crc kubenswrapper[4931]: I1125 22:01:39.673952 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"e502eeb6e6462755666e0a9f82f8ea7195e2305ad1bd997e1c4a8dc5d91dd65a"} Nov 25 22:01:39 crc kubenswrapper[4931]: I1125 22:01:39.674002 4931 scope.go:117] "RemoveContainer" containerID="8bda06aaa46440b32781ab98d8c14255214eac602bdadd7dbf0b48e5eb47636a" Nov 25 22:02:02 crc kubenswrapper[4931]: I1125 22:02:02.685562 4931 scope.go:117] "RemoveContainer" containerID="cf127098744248e8fdf2cf181ef5c685eb8d13b37e7279111ad03dd2b9aba597" Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.797247 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jjhc2"] Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.798628 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovn-controller" containerID="cri-o://55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c" gracePeriod=30 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.799121 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="sbdb" containerID="cri-o://5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34" gracePeriod=30 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.799186 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="nbdb" containerID="cri-o://a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca" gracePeriod=30 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.799244 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="northd" containerID="cri-o://ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d" gracePeriod=30 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.799296 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228" gracePeriod=30 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.799403 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kube-rbac-proxy-node" containerID="cri-o://364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d" gracePeriod=30 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.799470 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovn-acl-logging" containerID="cri-o://8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b" gracePeriod=30 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.838530 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" containerID="cri-o://560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9" gracePeriod=30 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.969357 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/3.log" Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.971973 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovn-acl-logging/0.log" Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.972704 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovn-controller/0.log" Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.973126 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228" exitCode=0 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.973145 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d" exitCode=0 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.973154 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b" exitCode=143 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.973163 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c" exitCode=143 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.973211 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228"} Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.973240 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d"} Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.973253 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b"} Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.973266 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c"} Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.977024 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/2.log" Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.977747 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/1.log" Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.977783 4931 generic.go:334] "Generic (PLEG): container finished" podID="ef281bd7-e1b1-41cd-b0a2-fa026829cf8f" containerID="18799a6d8db1c654581e5b8b65ff9892c6445ce545d95d10a08db229ac8c7296" exitCode=2 Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.977807 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xltbl" event={"ID":"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f","Type":"ContainerDied","Data":"18799a6d8db1c654581e5b8b65ff9892c6445ce545d95d10a08db229ac8c7296"} Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.977832 4931 scope.go:117] "RemoveContainer" containerID="0f0a4ef4d9180b2f077936dd456524894d3fa73180b7e2b21535745229e1af89" Nov 25 22:02:23 crc kubenswrapper[4931]: I1125 22:02:23.978380 4931 scope.go:117] "RemoveContainer" containerID="18799a6d8db1c654581e5b8b65ff9892c6445ce545d95d10a08db229ac8c7296" Nov 25 22:02:23 crc kubenswrapper[4931]: E1125 22:02:23.978572 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xltbl_openshift-multus(ef281bd7-e1b1-41cd-b0a2-fa026829cf8f)\"" pod="openshift-multus/multus-xltbl" podUID="ef281bd7-e1b1-41cd-b0a2-fa026829cf8f" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.138570 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/3.log" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.141259 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovn-acl-logging/0.log" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.141807 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovn-controller/0.log" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.142245 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203092 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-68v4l"] Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203386 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovn-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203407 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovn-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203427 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kubecfg-setup" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203440 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kubecfg-setup" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203456 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovn-acl-logging" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203467 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovn-acl-logging" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203482 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="sbdb" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203496 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="sbdb" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203508 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203519 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203532 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203543 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203559 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kube-rbac-proxy-node" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203571 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kube-rbac-proxy-node" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203591 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="northd" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203602 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="northd" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203619 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203631 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203646 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203657 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203670 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7" containerName="collect-profiles" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203681 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7" containerName="collect-profiles" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203696 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="nbdb" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203706 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="nbdb" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.203721 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203732 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203927 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203943 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="northd" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203961 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="nbdb" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203979 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.203994 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d7aa1a5-a5fd-4b48-97f9-0aa6eaed2ca7" containerName="collect-profiles" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.204006 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovn-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.204021 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.204038 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="sbdb" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.204053 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.204065 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="kube-rbac-proxy-node" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.204076 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovn-acl-logging" Nov 25 22:02:24 crc kubenswrapper[4931]: E1125 22:02:24.204269 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.204285 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.204436 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.204455 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerName="ovnkube-controller" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.206661 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302409 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nntdj\" (UniqueName: \"kubernetes.io/projected/b6be6a2b-7522-4af3-aa00-68fa970fd81c-kube-api-access-nntdj\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302455 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-config\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302476 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-env-overrides\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302499 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-etc-openvswitch\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302545 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-ovn\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302559 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-systemd\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302592 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-ovn-kubernetes\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302617 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-openvswitch\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302632 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-node-log\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302648 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-slash\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302667 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-bin\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302693 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-var-lib-openvswitch\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302714 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-netns\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302734 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302767 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovn-node-metrics-cert\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302791 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-script-lib\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302811 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-kubelet\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302841 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-systemd-units\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302905 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-log-socket\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.302936 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-netd\") pod \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\" (UID: \"b6be6a2b-7522-4af3-aa00-68fa970fd81c\") " Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303160 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303201 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303191 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303236 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303241 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303262 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303343 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303412 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303534 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303617 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303667 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303716 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-node-log" (OuterVolumeSpecName: "node-log") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303901 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-slash" (OuterVolumeSpecName: "host-slash") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303949 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.303975 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-log-socket" (OuterVolumeSpecName: "log-socket") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.304081 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.304848 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.309775 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6be6a2b-7522-4af3-aa00-68fa970fd81c-kube-api-access-nntdj" (OuterVolumeSpecName: "kube-api-access-nntdj") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "kube-api-access-nntdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.310951 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.325116 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "b6be6a2b-7522-4af3-aa00-68fa970fd81c" (UID: "b6be6a2b-7522-4af3-aa00-68fa970fd81c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.403705 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-var-lib-openvswitch\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.403759 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-etc-openvswitch\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.403781 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.403810 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a9982ae-8304-4072-bb2d-c855211daef3-ovnkube-config\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.403843 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-run-ovn\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.403947 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-slash\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404037 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a9982ae-8304-4072-bb2d-c855211daef3-env-overrides\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404102 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-cni-bin\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404185 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-run-netns\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404246 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-run-ovn-kubernetes\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404291 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp296\" (UniqueName: \"kubernetes.io/projected/2a9982ae-8304-4072-bb2d-c855211daef3-kube-api-access-jp296\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404317 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a9982ae-8304-4072-bb2d-c855211daef3-ovn-node-metrics-cert\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404344 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-cni-netd\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404376 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-log-socket\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404407 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-kubelet\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404462 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-run-openvswitch\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404494 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-systemd-units\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404520 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-run-systemd\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404543 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a9982ae-8304-4072-bb2d-c855211daef3-ovnkube-script-lib\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404571 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-node-log\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404637 4931 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404653 4931 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404669 4931 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404685 4931 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404701 4931 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404717 4931 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404737 4931 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404757 4931 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404769 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404780 4931 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404792 4931 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404805 4931 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404816 4931 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404828 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nntdj\" (UniqueName: \"kubernetes.io/projected/b6be6a2b-7522-4af3-aa00-68fa970fd81c-kube-api-access-nntdj\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404841 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404855 4931 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b6be6a2b-7522-4af3-aa00-68fa970fd81c-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404900 4931 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404914 4931 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404924 4931 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.404936 4931 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6be6a2b-7522-4af3-aa00-68fa970fd81c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506597 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-var-lib-openvswitch\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506674 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-etc-openvswitch\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506713 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506755 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a9982ae-8304-4072-bb2d-c855211daef3-ovnkube-config\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506799 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-run-ovn\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506822 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-slash\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506843 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a9982ae-8304-4072-bb2d-c855211daef3-env-overrides\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506868 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-cni-bin\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506920 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-run-netns\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506947 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-run-ovn-kubernetes\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506960 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.506987 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp296\" (UniqueName: \"kubernetes.io/projected/2a9982ae-8304-4072-bb2d-c855211daef3-kube-api-access-jp296\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507091 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a9982ae-8304-4072-bb2d-c855211daef3-ovn-node-metrics-cert\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507148 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-cni-netd\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507199 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-log-socket\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507236 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-kubelet\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507278 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-run-openvswitch\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507309 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-systemd-units\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507336 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-run-systemd\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507361 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a9982ae-8304-4072-bb2d-c855211daef3-ovnkube-script-lib\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507412 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-node-log\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507511 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-node-log\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507524 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-log-socket\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507559 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-kubelet\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507599 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-run-openvswitch\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507629 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-var-lib-openvswitch\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507642 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-systemd-units\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507678 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-etc-openvswitch\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.507685 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-run-systemd\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.508528 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-run-ovn\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.508657 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a9982ae-8304-4072-bb2d-c855211daef3-ovnkube-config\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.508670 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a9982ae-8304-4072-bb2d-c855211daef3-ovnkube-script-lib\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.508713 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-cni-netd\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.508758 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-cni-bin\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.508803 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-run-netns\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.508841 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-run-ovn-kubernetes\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.509085 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a9982ae-8304-4072-bb2d-c855211daef3-host-slash\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.509300 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a9982ae-8304-4072-bb2d-c855211daef3-env-overrides\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.511133 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a9982ae-8304-4072-bb2d-c855211daef3-ovn-node-metrics-cert\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.527298 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp296\" (UniqueName: \"kubernetes.io/projected/2a9982ae-8304-4072-bb2d-c855211daef3-kube-api-access-jp296\") pod \"ovnkube-node-68v4l\" (UID: \"2a9982ae-8304-4072-bb2d-c855211daef3\") " pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.827201 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.984464 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovnkube-controller/3.log" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.987530 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovn-acl-logging/0.log" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988094 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jjhc2_b6be6a2b-7522-4af3-aa00-68fa970fd81c/ovn-controller/0.log" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988574 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9" exitCode=0 Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988615 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34" exitCode=0 Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988627 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca" exitCode=0 Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988637 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" containerID="ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d" exitCode=0 Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988706 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9"} Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988727 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988741 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34"} Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988760 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca"} Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988773 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d"} Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988786 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jjhc2" event={"ID":"b6be6a2b-7522-4af3-aa00-68fa970fd81c","Type":"ContainerDied","Data":"5c01cc3149ce9ff29e4717d833639e7977a6755801166e0c7a511acf156515b0"} Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.988815 4931 scope.go:117] "RemoveContainer" containerID="560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9" Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.990861 4931 generic.go:334] "Generic (PLEG): container finished" podID="2a9982ae-8304-4072-bb2d-c855211daef3" containerID="a34b13d0691cb231bad6bb811e1acf425fd2b0574a81bece8e0457fb01bc00db" exitCode=0 Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.990947 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerDied","Data":"a34b13d0691cb231bad6bb811e1acf425fd2b0574a81bece8e0457fb01bc00db"} Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.990991 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerStarted","Data":"24a4a33b3ebabfb9481a6997a0a0b03ba23f4077b97a276118cb314a872bdafa"} Nov 25 22:02:24 crc kubenswrapper[4931]: I1125 22:02:24.995043 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/2.log" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.063912 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jjhc2"] Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.064355 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.068455 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jjhc2"] Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.080140 4931 scope.go:117] "RemoveContainer" containerID="5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.100657 4931 scope.go:117] "RemoveContainer" containerID="a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.116497 4931 scope.go:117] "RemoveContainer" containerID="ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.153353 4931 scope.go:117] "RemoveContainer" containerID="3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.174318 4931 scope.go:117] "RemoveContainer" containerID="364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.223919 4931 scope.go:117] "RemoveContainer" containerID="8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.245311 4931 scope.go:117] "RemoveContainer" containerID="55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.273408 4931 scope.go:117] "RemoveContainer" containerID="59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.292083 4931 scope.go:117] "RemoveContainer" containerID="560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.292534 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9\": container with ID starting with 560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9 not found: ID does not exist" containerID="560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.292579 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9"} err="failed to get container status \"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9\": rpc error: code = NotFound desc = could not find container \"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9\": container with ID starting with 560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.292610 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.293236 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\": container with ID starting with eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90 not found: ID does not exist" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.293278 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90"} err="failed to get container status \"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\": rpc error: code = NotFound desc = could not find container \"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\": container with ID starting with eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.293305 4931 scope.go:117] "RemoveContainer" containerID="5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.293587 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\": container with ID starting with 5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34 not found: ID does not exist" containerID="5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.293623 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34"} err="failed to get container status \"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\": rpc error: code = NotFound desc = could not find container \"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\": container with ID starting with 5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.293650 4931 scope.go:117] "RemoveContainer" containerID="a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.294066 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\": container with ID starting with a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca not found: ID does not exist" containerID="a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.294142 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca"} err="failed to get container status \"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\": rpc error: code = NotFound desc = could not find container \"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\": container with ID starting with a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.294211 4931 scope.go:117] "RemoveContainer" containerID="ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.294535 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\": container with ID starting with ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d not found: ID does not exist" containerID="ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.294573 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d"} err="failed to get container status \"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\": rpc error: code = NotFound desc = could not find container \"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\": container with ID starting with ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.294601 4931 scope.go:117] "RemoveContainer" containerID="3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.295057 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\": container with ID starting with 3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228 not found: ID does not exist" containerID="3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.295083 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228"} err="failed to get container status \"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\": rpc error: code = NotFound desc = could not find container \"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\": container with ID starting with 3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.295099 4931 scope.go:117] "RemoveContainer" containerID="364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.295557 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\": container with ID starting with 364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d not found: ID does not exist" containerID="364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.295612 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d"} err="failed to get container status \"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\": rpc error: code = NotFound desc = could not find container \"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\": container with ID starting with 364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.295646 4931 scope.go:117] "RemoveContainer" containerID="8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.296002 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\": container with ID starting with 8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b not found: ID does not exist" containerID="8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.296031 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b"} err="failed to get container status \"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\": rpc error: code = NotFound desc = could not find container \"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\": container with ID starting with 8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.296050 4931 scope.go:117] "RemoveContainer" containerID="55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.296306 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\": container with ID starting with 55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c not found: ID does not exist" containerID="55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.296342 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c"} err="failed to get container status \"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\": rpc error: code = NotFound desc = could not find container \"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\": container with ID starting with 55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.296369 4931 scope.go:117] "RemoveContainer" containerID="59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98" Nov 25 22:02:25 crc kubenswrapper[4931]: E1125 22:02:25.296705 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\": container with ID starting with 59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98 not found: ID does not exist" containerID="59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.296739 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98"} err="failed to get container status \"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\": rpc error: code = NotFound desc = could not find container \"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\": container with ID starting with 59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.296761 4931 scope.go:117] "RemoveContainer" containerID="560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.297223 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9"} err="failed to get container status \"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9\": rpc error: code = NotFound desc = could not find container \"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9\": container with ID starting with 560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.297247 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.297683 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90"} err="failed to get container status \"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\": rpc error: code = NotFound desc = could not find container \"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\": container with ID starting with eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.297715 4931 scope.go:117] "RemoveContainer" containerID="5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.298106 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34"} err="failed to get container status \"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\": rpc error: code = NotFound desc = could not find container \"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\": container with ID starting with 5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.298141 4931 scope.go:117] "RemoveContainer" containerID="a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.298409 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca"} err="failed to get container status \"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\": rpc error: code = NotFound desc = could not find container \"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\": container with ID starting with a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.298432 4931 scope.go:117] "RemoveContainer" containerID="ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.298697 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d"} err="failed to get container status \"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\": rpc error: code = NotFound desc = could not find container \"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\": container with ID starting with ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.298723 4931 scope.go:117] "RemoveContainer" containerID="3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.298966 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228"} err="failed to get container status \"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\": rpc error: code = NotFound desc = could not find container \"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\": container with ID starting with 3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.298988 4931 scope.go:117] "RemoveContainer" containerID="364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.299176 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d"} err="failed to get container status \"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\": rpc error: code = NotFound desc = could not find container \"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\": container with ID starting with 364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.299196 4931 scope.go:117] "RemoveContainer" containerID="8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.299405 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b"} err="failed to get container status \"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\": rpc error: code = NotFound desc = could not find container \"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\": container with ID starting with 8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.299437 4931 scope.go:117] "RemoveContainer" containerID="55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.299679 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c"} err="failed to get container status \"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\": rpc error: code = NotFound desc = could not find container \"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\": container with ID starting with 55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.299701 4931 scope.go:117] "RemoveContainer" containerID="59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.299941 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98"} err="failed to get container status \"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\": rpc error: code = NotFound desc = could not find container \"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\": container with ID starting with 59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.299972 4931 scope.go:117] "RemoveContainer" containerID="560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.300537 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9"} err="failed to get container status \"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9\": rpc error: code = NotFound desc = could not find container \"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9\": container with ID starting with 560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.300559 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.300977 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90"} err="failed to get container status \"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\": rpc error: code = NotFound desc = could not find container \"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\": container with ID starting with eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.301007 4931 scope.go:117] "RemoveContainer" containerID="5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.301263 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34"} err="failed to get container status \"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\": rpc error: code = NotFound desc = could not find container \"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\": container with ID starting with 5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.301286 4931 scope.go:117] "RemoveContainer" containerID="a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.301676 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca"} err="failed to get container status \"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\": rpc error: code = NotFound desc = could not find container \"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\": container with ID starting with a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.301701 4931 scope.go:117] "RemoveContainer" containerID="ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.302305 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d"} err="failed to get container status \"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\": rpc error: code = NotFound desc = could not find container \"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\": container with ID starting with ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.302332 4931 scope.go:117] "RemoveContainer" containerID="3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.304281 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228"} err="failed to get container status \"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\": rpc error: code = NotFound desc = could not find container \"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\": container with ID starting with 3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.304315 4931 scope.go:117] "RemoveContainer" containerID="364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.304810 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d"} err="failed to get container status \"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\": rpc error: code = NotFound desc = could not find container \"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\": container with ID starting with 364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.304832 4931 scope.go:117] "RemoveContainer" containerID="8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.305103 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b"} err="failed to get container status \"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\": rpc error: code = NotFound desc = could not find container \"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\": container with ID starting with 8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.305122 4931 scope.go:117] "RemoveContainer" containerID="55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.305537 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c"} err="failed to get container status \"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\": rpc error: code = NotFound desc = could not find container \"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\": container with ID starting with 55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.305565 4931 scope.go:117] "RemoveContainer" containerID="59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.305781 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98"} err="failed to get container status \"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\": rpc error: code = NotFound desc = could not find container \"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\": container with ID starting with 59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.305798 4931 scope.go:117] "RemoveContainer" containerID="560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.306217 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9"} err="failed to get container status \"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9\": rpc error: code = NotFound desc = could not find container \"560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9\": container with ID starting with 560ebdd06b3629e0030fda26416ee1e42247dc6f2d1bd2440ee92f87d72ecce9 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.306243 4931 scope.go:117] "RemoveContainer" containerID="eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.306477 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90"} err="failed to get container status \"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\": rpc error: code = NotFound desc = could not find container \"eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90\": container with ID starting with eebfb37a968589d98f863ab41a2d4c3f505bae2b1df8b80c31e3e7021f0a9e90 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.306511 4931 scope.go:117] "RemoveContainer" containerID="5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.306760 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34"} err="failed to get container status \"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\": rpc error: code = NotFound desc = could not find container \"5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34\": container with ID starting with 5910619471d1ba33ec597288349548cfabeb6572fad13519b54ca1aed5bceb34 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.306781 4931 scope.go:117] "RemoveContainer" containerID="a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.307004 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca"} err="failed to get container status \"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\": rpc error: code = NotFound desc = could not find container \"a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca\": container with ID starting with a02f3be0b2ff86d9eaf8868da07b45855bfff602c6b372b704d461b1d83363ca not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.307021 4931 scope.go:117] "RemoveContainer" containerID="ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.307246 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d"} err="failed to get container status \"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\": rpc error: code = NotFound desc = could not find container \"ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d\": container with ID starting with ef2b79a3d31b58835fa1bfef7d4eb85f1f03ba6e49d1bd338b4fea0c7b3cf56d not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.307272 4931 scope.go:117] "RemoveContainer" containerID="3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.307527 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228"} err="failed to get container status \"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\": rpc error: code = NotFound desc = could not find container \"3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228\": container with ID starting with 3301fc1b351c195769e7ab178a15acfda4bd30c587828301003932856ed2b228 not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.307550 4931 scope.go:117] "RemoveContainer" containerID="364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.307735 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d"} err="failed to get container status \"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\": rpc error: code = NotFound desc = could not find container \"364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d\": container with ID starting with 364c661d44c2d68b3a151dc31c24905875d3d93eb7366eb335c9bbe7da88783d not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.307751 4931 scope.go:117] "RemoveContainer" containerID="8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.308157 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b"} err="failed to get container status \"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\": rpc error: code = NotFound desc = could not find container \"8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b\": container with ID starting with 8d9d2bea7c74901a25d0b75af4c8bb33721c5c32fd2d5f1632d6fd62e7bec69b not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.308183 4931 scope.go:117] "RemoveContainer" containerID="55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.308564 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c"} err="failed to get container status \"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\": rpc error: code = NotFound desc = could not find container \"55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c\": container with ID starting with 55d53987b7869984ddd36f96481401e95132303e4f3b3322626a4971fc8f0f0c not found: ID does not exist" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.308651 4931 scope.go:117] "RemoveContainer" containerID="59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98" Nov 25 22:02:25 crc kubenswrapper[4931]: I1125 22:02:25.309001 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98"} err="failed to get container status \"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\": rpc error: code = NotFound desc = could not find container \"59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98\": container with ID starting with 59784377925dd4a07d7f0df1a314c4c9a3f1f71cd5f7de4cc5956f877f16ca98 not found: ID does not exist" Nov 25 22:02:26 crc kubenswrapper[4931]: I1125 22:02:26.009725 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerStarted","Data":"50f3d478f4400fcce91a796955c679ebc698b13b1efa10311713b3ce41e07e2d"} Nov 25 22:02:26 crc kubenswrapper[4931]: I1125 22:02:26.010099 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerStarted","Data":"98792f0e7b268dd49c9682e7e24104135f2bc4a1f522a23279650d205b5b7070"} Nov 25 22:02:26 crc kubenswrapper[4931]: I1125 22:02:26.010156 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerStarted","Data":"e73890d25c2f07e8dad42df91ae198e5d53210ff558eb2cf615c7923afd0931f"} Nov 25 22:02:26 crc kubenswrapper[4931]: I1125 22:02:26.010177 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerStarted","Data":"9e8db9658b06cbff737aa2aeb44d5be4e5132c777d7823b9ee0dc18dea8f908b"} Nov 25 22:02:26 crc kubenswrapper[4931]: I1125 22:02:26.010194 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerStarted","Data":"fc1decded9858afe21a22c36bee443a76f70c58fc2ac49e41ec658d7492dd45b"} Nov 25 22:02:26 crc kubenswrapper[4931]: I1125 22:02:26.010214 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerStarted","Data":"d58c487b7bc349b69f62413859da53056e91d3a36c9e9f767d27cc9411b97d89"} Nov 25 22:02:26 crc kubenswrapper[4931]: I1125 22:02:26.431723 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6be6a2b-7522-4af3-aa00-68fa970fd81c" path="/var/lib/kubelet/pods/b6be6a2b-7522-4af3-aa00-68fa970fd81c/volumes" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.077705 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j"] Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.079748 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.082612 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.254694 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w9fc\" (UniqueName: \"kubernetes.io/projected/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-kube-api-access-6w9fc\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.254766 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.254806 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.356091 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.356191 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.356272 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w9fc\" (UniqueName: \"kubernetes.io/projected/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-kube-api-access-6w9fc\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.357612 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.358423 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.386460 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w9fc\" (UniqueName: \"kubernetes.io/projected/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-kube-api-access-6w9fc\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: I1125 22:02:28.395568 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: E1125 22:02:28.440179 4931 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(71234e93925bac04f9d43da072c225460238de9b6aea1c7e1014cf387fe3ffb6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 22:02:28 crc kubenswrapper[4931]: E1125 22:02:28.440287 4931 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(71234e93925bac04f9d43da072c225460238de9b6aea1c7e1014cf387fe3ffb6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: E1125 22:02:28.440325 4931 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(71234e93925bac04f9d43da072c225460238de9b6aea1c7e1014cf387fe3ffb6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:28 crc kubenswrapper[4931]: E1125 22:02:28.440406 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace(79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace(79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(71234e93925bac04f9d43da072c225460238de9b6aea1c7e1014cf387fe3ffb6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" Nov 25 22:02:29 crc kubenswrapper[4931]: I1125 22:02:29.038052 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerStarted","Data":"f8d53f30dbec891e08a648e4a41e06d8b083e885155db1fd75d6a253fc9e022b"} Nov 25 22:02:31 crc kubenswrapper[4931]: I1125 22:02:31.056645 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" event={"ID":"2a9982ae-8304-4072-bb2d-c855211daef3","Type":"ContainerStarted","Data":"a5e5ea2c1742bbc2e44543bac16bd67abf28ebb148259dacd657dc84f349d7e9"} Nov 25 22:02:31 crc kubenswrapper[4931]: I1125 22:02:31.057248 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:31 crc kubenswrapper[4931]: I1125 22:02:31.057268 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:31 crc kubenswrapper[4931]: I1125 22:02:31.067474 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j"] Nov 25 22:02:31 crc kubenswrapper[4931]: I1125 22:02:31.067628 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:31 crc kubenswrapper[4931]: I1125 22:02:31.068170 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:31 crc kubenswrapper[4931]: I1125 22:02:31.090005 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" podStartSLOduration=7.089988776 podStartE2EDuration="7.089988776s" podCreationTimestamp="2025-11-25 22:02:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:02:31.088250428 +0000 UTC m=+689.444762547" watchObservedRunningTime="2025-11-25 22:02:31.089988776 +0000 UTC m=+689.446500895" Nov 25 22:02:31 crc kubenswrapper[4931]: E1125 22:02:31.096730 4931 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(d58c0b688ccaeec47d34b457b7a74a83c40008d7f301307a60bdbbd39c2ea6b0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 22:02:31 crc kubenswrapper[4931]: E1125 22:02:31.096784 4931 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(d58c0b688ccaeec47d34b457b7a74a83c40008d7f301307a60bdbbd39c2ea6b0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:31 crc kubenswrapper[4931]: E1125 22:02:31.096803 4931 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(d58c0b688ccaeec47d34b457b7a74a83c40008d7f301307a60bdbbd39c2ea6b0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:31 crc kubenswrapper[4931]: E1125 22:02:31.096844 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace(79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace(79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(d58c0b688ccaeec47d34b457b7a74a83c40008d7f301307a60bdbbd39c2ea6b0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" Nov 25 22:02:31 crc kubenswrapper[4931]: I1125 22:02:31.098587 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:32 crc kubenswrapper[4931]: I1125 22:02:32.066317 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:32 crc kubenswrapper[4931]: I1125 22:02:32.110570 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:35 crc kubenswrapper[4931]: I1125 22:02:35.423295 4931 scope.go:117] "RemoveContainer" containerID="18799a6d8db1c654581e5b8b65ff9892c6445ce545d95d10a08db229ac8c7296" Nov 25 22:02:35 crc kubenswrapper[4931]: E1125 22:02:35.424063 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xltbl_openshift-multus(ef281bd7-e1b1-41cd-b0a2-fa026829cf8f)\"" pod="openshift-multus/multus-xltbl" podUID="ef281bd7-e1b1-41cd-b0a2-fa026829cf8f" Nov 25 22:02:44 crc kubenswrapper[4931]: I1125 22:02:44.421947 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:44 crc kubenswrapper[4931]: I1125 22:02:44.423557 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:44 crc kubenswrapper[4931]: E1125 22:02:44.463274 4931 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(9ef6fc36ae27c4e8669f0f30e97345d84bacddc85edd448d41aeb1b0eda9f05a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 22:02:44 crc kubenswrapper[4931]: E1125 22:02:44.463666 4931 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(9ef6fc36ae27c4e8669f0f30e97345d84bacddc85edd448d41aeb1b0eda9f05a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:44 crc kubenswrapper[4931]: E1125 22:02:44.463705 4931 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(9ef6fc36ae27c4e8669f0f30e97345d84bacddc85edd448d41aeb1b0eda9f05a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:44 crc kubenswrapper[4931]: E1125 22:02:44.463791 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace(79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace(79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_openshift-marketplace_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94_0(9ef6fc36ae27c4e8669f0f30e97345d84bacddc85edd448d41aeb1b0eda9f05a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" Nov 25 22:02:46 crc kubenswrapper[4931]: I1125 22:02:46.423585 4931 scope.go:117] "RemoveContainer" containerID="18799a6d8db1c654581e5b8b65ff9892c6445ce545d95d10a08db229ac8c7296" Nov 25 22:02:47 crc kubenswrapper[4931]: I1125 22:02:47.187144 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xltbl_ef281bd7-e1b1-41cd-b0a2-fa026829cf8f/kube-multus/2.log" Nov 25 22:02:47 crc kubenswrapper[4931]: I1125 22:02:47.187588 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xltbl" event={"ID":"ef281bd7-e1b1-41cd-b0a2-fa026829cf8f","Type":"ContainerStarted","Data":"2a9d7b3ee6537245a195a604fa449d13ae55ed776aa7a5d3f752793aae67dc99"} Nov 25 22:02:54 crc kubenswrapper[4931]: I1125 22:02:54.866926 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-68v4l" Nov 25 22:02:57 crc kubenswrapper[4931]: I1125 22:02:57.421996 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:57 crc kubenswrapper[4931]: I1125 22:02:57.422999 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:02:57 crc kubenswrapper[4931]: I1125 22:02:57.688724 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j"] Nov 25 22:02:57 crc kubenswrapper[4931]: W1125 22:02:57.701390 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79cfa2ed_ad95_45eb_a55d_3f1fb6f08e94.slice/crio-e55302c4ee7da69f6c4f630026cadb02ec2e2aa94fd6bb985cf23c62192b9d41 WatchSource:0}: Error finding container e55302c4ee7da69f6c4f630026cadb02ec2e2aa94fd6bb985cf23c62192b9d41: Status 404 returned error can't find the container with id e55302c4ee7da69f6c4f630026cadb02ec2e2aa94fd6bb985cf23c62192b9d41 Nov 25 22:02:58 crc kubenswrapper[4931]: I1125 22:02:58.262791 4931 generic.go:334] "Generic (PLEG): container finished" podID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerID="4b54774a1aba3466512d8a3557cbc17a97fb62da9240129ce23c34bae60e7712" exitCode=0 Nov 25 22:02:58 crc kubenswrapper[4931]: I1125 22:02:58.262949 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" event={"ID":"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94","Type":"ContainerDied","Data":"4b54774a1aba3466512d8a3557cbc17a97fb62da9240129ce23c34bae60e7712"} Nov 25 22:02:58 crc kubenswrapper[4931]: I1125 22:02:58.262996 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" event={"ID":"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94","Type":"ContainerStarted","Data":"e55302c4ee7da69f6c4f630026cadb02ec2e2aa94fd6bb985cf23c62192b9d41"} Nov 25 22:02:58 crc kubenswrapper[4931]: I1125 22:02:58.268000 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 22:03:00 crc kubenswrapper[4931]: I1125 22:03:00.277773 4931 generic.go:334] "Generic (PLEG): container finished" podID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerID="466a75a3fd979ccf56d853d093ffd1a1f61a646ddac1c1cb7e76ef4921fd8b79" exitCode=0 Nov 25 22:03:00 crc kubenswrapper[4931]: I1125 22:03:00.278179 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" event={"ID":"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94","Type":"ContainerDied","Data":"466a75a3fd979ccf56d853d093ffd1a1f61a646ddac1c1cb7e76ef4921fd8b79"} Nov 25 22:03:01 crc kubenswrapper[4931]: I1125 22:03:01.287855 4931 generic.go:334] "Generic (PLEG): container finished" podID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerID="286e83fccf69f0ed2362d0c53ccc9405ae6a77d326db3c25d6e260628086a11b" exitCode=0 Nov 25 22:03:01 crc kubenswrapper[4931]: I1125 22:03:01.287950 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" event={"ID":"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94","Type":"ContainerDied","Data":"286e83fccf69f0ed2362d0c53ccc9405ae6a77d326db3c25d6e260628086a11b"} Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.667024 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.776204 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-bundle\") pod \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.776265 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w9fc\" (UniqueName: \"kubernetes.io/projected/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-kube-api-access-6w9fc\") pod \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.776303 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-util\") pod \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\" (UID: \"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94\") " Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.777923 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-bundle" (OuterVolumeSpecName: "bundle") pod "79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" (UID: "79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.783353 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-kube-api-access-6w9fc" (OuterVolumeSpecName: "kube-api-access-6w9fc") pod "79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" (UID: "79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94"). InnerVolumeSpecName "kube-api-access-6w9fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.808645 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-util" (OuterVolumeSpecName: "util") pod "79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" (UID: "79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.877694 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.877770 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w9fc\" (UniqueName: \"kubernetes.io/projected/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-kube-api-access-6w9fc\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:02 crc kubenswrapper[4931]: I1125 22:03:02.877790 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94-util\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:03 crc kubenswrapper[4931]: I1125 22:03:03.303525 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" event={"ID":"79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94","Type":"ContainerDied","Data":"e55302c4ee7da69f6c4f630026cadb02ec2e2aa94fd6bb985cf23c62192b9d41"} Nov 25 22:03:03 crc kubenswrapper[4931]: I1125 22:03:03.303583 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e55302c4ee7da69f6c4f630026cadb02ec2e2aa94fd6bb985cf23c62192b9d41" Nov 25 22:03:03 crc kubenswrapper[4931]: I1125 22:03:03.303587 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.326698 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz"] Nov 25 22:03:17 crc kubenswrapper[4931]: E1125 22:03:17.327507 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerName="extract" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.327521 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerName="extract" Nov 25 22:03:17 crc kubenswrapper[4931]: E1125 22:03:17.327529 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerName="pull" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.327535 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerName="pull" Nov 25 22:03:17 crc kubenswrapper[4931]: E1125 22:03:17.327547 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerName="util" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.327554 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerName="util" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.327633 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94" containerName="extract" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.327996 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.332709 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.333038 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.333257 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.333354 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-44wj8" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.337079 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.372913 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60182b4c-52b4-47a9-8c5a-ab0274136ef1-webhook-cert\") pod \"metallb-operator-controller-manager-5bfcf68b5b-79jjz\" (UID: \"60182b4c-52b4-47a9-8c5a-ab0274136ef1\") " pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.372963 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60182b4c-52b4-47a9-8c5a-ab0274136ef1-apiservice-cert\") pod \"metallb-operator-controller-manager-5bfcf68b5b-79jjz\" (UID: \"60182b4c-52b4-47a9-8c5a-ab0274136ef1\") " pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.373021 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j79j8\" (UniqueName: \"kubernetes.io/projected/60182b4c-52b4-47a9-8c5a-ab0274136ef1-kube-api-access-j79j8\") pod \"metallb-operator-controller-manager-5bfcf68b5b-79jjz\" (UID: \"60182b4c-52b4-47a9-8c5a-ab0274136ef1\") " pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.378702 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz"] Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.474176 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60182b4c-52b4-47a9-8c5a-ab0274136ef1-webhook-cert\") pod \"metallb-operator-controller-manager-5bfcf68b5b-79jjz\" (UID: \"60182b4c-52b4-47a9-8c5a-ab0274136ef1\") " pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.474219 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60182b4c-52b4-47a9-8c5a-ab0274136ef1-apiservice-cert\") pod \"metallb-operator-controller-manager-5bfcf68b5b-79jjz\" (UID: \"60182b4c-52b4-47a9-8c5a-ab0274136ef1\") " pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.474280 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j79j8\" (UniqueName: \"kubernetes.io/projected/60182b4c-52b4-47a9-8c5a-ab0274136ef1-kube-api-access-j79j8\") pod \"metallb-operator-controller-manager-5bfcf68b5b-79jjz\" (UID: \"60182b4c-52b4-47a9-8c5a-ab0274136ef1\") " pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.480464 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60182b4c-52b4-47a9-8c5a-ab0274136ef1-apiservice-cert\") pod \"metallb-operator-controller-manager-5bfcf68b5b-79jjz\" (UID: \"60182b4c-52b4-47a9-8c5a-ab0274136ef1\") " pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.487159 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60182b4c-52b4-47a9-8c5a-ab0274136ef1-webhook-cert\") pod \"metallb-operator-controller-manager-5bfcf68b5b-79jjz\" (UID: \"60182b4c-52b4-47a9-8c5a-ab0274136ef1\") " pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.488749 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j79j8\" (UniqueName: \"kubernetes.io/projected/60182b4c-52b4-47a9-8c5a-ab0274136ef1-kube-api-access-j79j8\") pod \"metallb-operator-controller-manager-5bfcf68b5b-79jjz\" (UID: \"60182b4c-52b4-47a9-8c5a-ab0274136ef1\") " pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.551946 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn"] Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.553170 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.562020 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.565325 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-9qv82" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.566047 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.575205 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ba8977d1-518a-46c8-b174-073b37d6ebb6-webhook-cert\") pod \"metallb-operator-webhook-server-6dcc996cd8-f5vhn\" (UID: \"ba8977d1-518a-46c8-b174-073b37d6ebb6\") " pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.575260 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5qzc\" (UniqueName: \"kubernetes.io/projected/ba8977d1-518a-46c8-b174-073b37d6ebb6-kube-api-access-z5qzc\") pod \"metallb-operator-webhook-server-6dcc996cd8-f5vhn\" (UID: \"ba8977d1-518a-46c8-b174-073b37d6ebb6\") " pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.575378 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ba8977d1-518a-46c8-b174-073b37d6ebb6-apiservice-cert\") pod \"metallb-operator-webhook-server-6dcc996cd8-f5vhn\" (UID: \"ba8977d1-518a-46c8-b174-073b37d6ebb6\") " pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.578273 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn"] Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.644435 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.677489 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ba8977d1-518a-46c8-b174-073b37d6ebb6-apiservice-cert\") pod \"metallb-operator-webhook-server-6dcc996cd8-f5vhn\" (UID: \"ba8977d1-518a-46c8-b174-073b37d6ebb6\") " pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.677567 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ba8977d1-518a-46c8-b174-073b37d6ebb6-webhook-cert\") pod \"metallb-operator-webhook-server-6dcc996cd8-f5vhn\" (UID: \"ba8977d1-518a-46c8-b174-073b37d6ebb6\") " pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.677606 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5qzc\" (UniqueName: \"kubernetes.io/projected/ba8977d1-518a-46c8-b174-073b37d6ebb6-kube-api-access-z5qzc\") pod \"metallb-operator-webhook-server-6dcc996cd8-f5vhn\" (UID: \"ba8977d1-518a-46c8-b174-073b37d6ebb6\") " pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.686856 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ba8977d1-518a-46c8-b174-073b37d6ebb6-apiservice-cert\") pod \"metallb-operator-webhook-server-6dcc996cd8-f5vhn\" (UID: \"ba8977d1-518a-46c8-b174-073b37d6ebb6\") " pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.688455 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ba8977d1-518a-46c8-b174-073b37d6ebb6-webhook-cert\") pod \"metallb-operator-webhook-server-6dcc996cd8-f5vhn\" (UID: \"ba8977d1-518a-46c8-b174-073b37d6ebb6\") " pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.706155 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5qzc\" (UniqueName: \"kubernetes.io/projected/ba8977d1-518a-46c8-b174-073b37d6ebb6-kube-api-access-z5qzc\") pod \"metallb-operator-webhook-server-6dcc996cd8-f5vhn\" (UID: \"ba8977d1-518a-46c8-b174-073b37d6ebb6\") " pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.878584 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:17 crc kubenswrapper[4931]: I1125 22:03:17.880473 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz"] Nov 25 22:03:17 crc kubenswrapper[4931]: W1125 22:03:17.893733 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60182b4c_52b4_47a9_8c5a_ab0274136ef1.slice/crio-3c777664be84b58005e7c9019e53dd3b6454f2e458b1c0a7a5a6e1a0df1191b8 WatchSource:0}: Error finding container 3c777664be84b58005e7c9019e53dd3b6454f2e458b1c0a7a5a6e1a0df1191b8: Status 404 returned error can't find the container with id 3c777664be84b58005e7c9019e53dd3b6454f2e458b1c0a7a5a6e1a0df1191b8 Nov 25 22:03:18 crc kubenswrapper[4931]: I1125 22:03:18.089843 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn"] Nov 25 22:03:18 crc kubenswrapper[4931]: W1125 22:03:18.095089 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba8977d1_518a_46c8_b174_073b37d6ebb6.slice/crio-3905c2084fb39de516343a6573ebf50a38c0b908c6b3c3efff285d5862236655 WatchSource:0}: Error finding container 3905c2084fb39de516343a6573ebf50a38c0b908c6b3c3efff285d5862236655: Status 404 returned error can't find the container with id 3905c2084fb39de516343a6573ebf50a38c0b908c6b3c3efff285d5862236655 Nov 25 22:03:18 crc kubenswrapper[4931]: I1125 22:03:18.404433 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" event={"ID":"60182b4c-52b4-47a9-8c5a-ab0274136ef1","Type":"ContainerStarted","Data":"3c777664be84b58005e7c9019e53dd3b6454f2e458b1c0a7a5a6e1a0df1191b8"} Nov 25 22:03:18 crc kubenswrapper[4931]: I1125 22:03:18.405511 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" event={"ID":"ba8977d1-518a-46c8-b174-073b37d6ebb6","Type":"ContainerStarted","Data":"3905c2084fb39de516343a6573ebf50a38c0b908c6b3c3efff285d5862236655"} Nov 25 22:03:21 crc kubenswrapper[4931]: I1125 22:03:21.431704 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" event={"ID":"60182b4c-52b4-47a9-8c5a-ab0274136ef1","Type":"ContainerStarted","Data":"a00ec071c6e2988ead68259cb195dfb80083380991f1ec8272588984bd389df1"} Nov 25 22:03:21 crc kubenswrapper[4931]: I1125 22:03:21.432402 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:21 crc kubenswrapper[4931]: I1125 22:03:21.457205 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" podStartSLOduration=1.385492167 podStartE2EDuration="4.457184332s" podCreationTimestamp="2025-11-25 22:03:17 +0000 UTC" firstStartedPulling="2025-11-25 22:03:17.896505502 +0000 UTC m=+736.253017621" lastFinishedPulling="2025-11-25 22:03:20.968197667 +0000 UTC m=+739.324709786" observedRunningTime="2025-11-25 22:03:21.454314442 +0000 UTC m=+739.810826581" watchObservedRunningTime="2025-11-25 22:03:21.457184332 +0000 UTC m=+739.813696461" Nov 25 22:03:23 crc kubenswrapper[4931]: I1125 22:03:23.443554 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" event={"ID":"ba8977d1-518a-46c8-b174-073b37d6ebb6","Type":"ContainerStarted","Data":"0ddebcc321f261f0d4517a6bfda15d8fb85a08486663e77d4742ddf8c98d1299"} Nov 25 22:03:23 crc kubenswrapper[4931]: I1125 22:03:23.443922 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:23 crc kubenswrapper[4931]: I1125 22:03:23.480500 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" podStartSLOduration=1.877248687 podStartE2EDuration="6.480471701s" podCreationTimestamp="2025-11-25 22:03:17 +0000 UTC" firstStartedPulling="2025-11-25 22:03:18.096928815 +0000 UTC m=+736.453440944" lastFinishedPulling="2025-11-25 22:03:22.700151819 +0000 UTC m=+741.056663958" observedRunningTime="2025-11-25 22:03:23.475785381 +0000 UTC m=+741.832297530" watchObservedRunningTime="2025-11-25 22:03:23.480471701 +0000 UTC m=+741.836983850" Nov 25 22:03:37 crc kubenswrapper[4931]: I1125 22:03:37.883844 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6dcc996cd8-f5vhn" Nov 25 22:03:41 crc kubenswrapper[4931]: I1125 22:03:41.886020 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kvz7t"] Nov 25 22:03:41 crc kubenswrapper[4931]: I1125 22:03:41.886826 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" podUID="74cbb6cf-4717-439c-b712-67718c154742" containerName="controller-manager" containerID="cri-o://98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd" gracePeriod=30 Nov 25 22:03:41 crc kubenswrapper[4931]: I1125 22:03:41.968492 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st"] Nov 25 22:03:41 crc kubenswrapper[4931]: I1125 22:03:41.968654 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" podUID="fada681c-6ade-49e4-b324-1985853bc61e" containerName="route-controller-manager" containerID="cri-o://7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753" gracePeriod=30 Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.226774 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.283475 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.395625 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x2px\" (UniqueName: \"kubernetes.io/projected/fada681c-6ade-49e4-b324-1985853bc61e-kube-api-access-8x2px\") pod \"fada681c-6ade-49e4-b324-1985853bc61e\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.395680 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fada681c-6ade-49e4-b324-1985853bc61e-serving-cert\") pod \"fada681c-6ade-49e4-b324-1985853bc61e\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.395720 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-config\") pod \"fada681c-6ade-49e4-b324-1985853bc61e\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.395765 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-client-ca\") pod \"74cbb6cf-4717-439c-b712-67718c154742\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.395780 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-proxy-ca-bundles\") pod \"74cbb6cf-4717-439c-b712-67718c154742\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.395796 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-config\") pod \"74cbb6cf-4717-439c-b712-67718c154742\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.395813 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-client-ca\") pod \"fada681c-6ade-49e4-b324-1985853bc61e\" (UID: \"fada681c-6ade-49e4-b324-1985853bc61e\") " Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.395830 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74cbb6cf-4717-439c-b712-67718c154742-serving-cert\") pod \"74cbb6cf-4717-439c-b712-67718c154742\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.395895 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srckn\" (UniqueName: \"kubernetes.io/projected/74cbb6cf-4717-439c-b712-67718c154742-kube-api-access-srckn\") pod \"74cbb6cf-4717-439c-b712-67718c154742\" (UID: \"74cbb6cf-4717-439c-b712-67718c154742\") " Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.396491 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-client-ca" (OuterVolumeSpecName: "client-ca") pod "fada681c-6ade-49e4-b324-1985853bc61e" (UID: "fada681c-6ade-49e4-b324-1985853bc61e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.396550 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-client-ca" (OuterVolumeSpecName: "client-ca") pod "74cbb6cf-4717-439c-b712-67718c154742" (UID: "74cbb6cf-4717-439c-b712-67718c154742"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.396583 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "74cbb6cf-4717-439c-b712-67718c154742" (UID: "74cbb6cf-4717-439c-b712-67718c154742"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.396691 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-config" (OuterVolumeSpecName: "config") pod "fada681c-6ade-49e4-b324-1985853bc61e" (UID: "fada681c-6ade-49e4-b324-1985853bc61e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.397267 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-config" (OuterVolumeSpecName: "config") pod "74cbb6cf-4717-439c-b712-67718c154742" (UID: "74cbb6cf-4717-439c-b712-67718c154742"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.401324 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74cbb6cf-4717-439c-b712-67718c154742-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "74cbb6cf-4717-439c-b712-67718c154742" (UID: "74cbb6cf-4717-439c-b712-67718c154742"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.401812 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74cbb6cf-4717-439c-b712-67718c154742-kube-api-access-srckn" (OuterVolumeSpecName: "kube-api-access-srckn") pod "74cbb6cf-4717-439c-b712-67718c154742" (UID: "74cbb6cf-4717-439c-b712-67718c154742"). InnerVolumeSpecName "kube-api-access-srckn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.404998 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fada681c-6ade-49e4-b324-1985853bc61e-kube-api-access-8x2px" (OuterVolumeSpecName: "kube-api-access-8x2px") pod "fada681c-6ade-49e4-b324-1985853bc61e" (UID: "fada681c-6ade-49e4-b324-1985853bc61e"). InnerVolumeSpecName "kube-api-access-8x2px". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.406834 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fada681c-6ade-49e4-b324-1985853bc61e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fada681c-6ade-49e4-b324-1985853bc61e" (UID: "fada681c-6ade-49e4-b324-1985853bc61e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.497823 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x2px\" (UniqueName: \"kubernetes.io/projected/fada681c-6ade-49e4-b324-1985853bc61e-kube-api-access-8x2px\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.497905 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fada681c-6ade-49e4-b324-1985853bc61e-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.497928 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.497947 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.497966 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.497986 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fada681c-6ade-49e4-b324-1985853bc61e-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.498004 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cbb6cf-4717-439c-b712-67718c154742-config\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.498022 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74cbb6cf-4717-439c-b712-67718c154742-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.498041 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srckn\" (UniqueName: \"kubernetes.io/projected/74cbb6cf-4717-439c-b712-67718c154742-kube-api-access-srckn\") on node \"crc\" DevicePath \"\"" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.549538 4931 generic.go:334] "Generic (PLEG): container finished" podID="74cbb6cf-4717-439c-b712-67718c154742" containerID="98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd" exitCode=0 Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.549614 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" event={"ID":"74cbb6cf-4717-439c-b712-67718c154742","Type":"ContainerDied","Data":"98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd"} Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.549643 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" event={"ID":"74cbb6cf-4717-439c-b712-67718c154742","Type":"ContainerDied","Data":"65181b64fc25f693bd5d81c886a8293a876f472850faf7995121ecaa6c7b1942"} Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.549647 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kvz7t" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.549660 4931 scope.go:117] "RemoveContainer" containerID="98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.553635 4931 generic.go:334] "Generic (PLEG): container finished" podID="fada681c-6ade-49e4-b324-1985853bc61e" containerID="7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753" exitCode=0 Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.553665 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" event={"ID":"fada681c-6ade-49e4-b324-1985853bc61e","Type":"ContainerDied","Data":"7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753"} Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.553679 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.553708 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st" event={"ID":"fada681c-6ade-49e4-b324-1985853bc61e","Type":"ContainerDied","Data":"e07c9a9d080ea1c3fd5e19dd6f5e4bebdc621177a01e5c9fd3d9d377bac21c13"} Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.568405 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kvz7t"] Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.573747 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kvz7t"] Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.576714 4931 scope.go:117] "RemoveContainer" containerID="98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.582604 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st"] Nov 25 22:03:42 crc kubenswrapper[4931]: E1125 22:03:42.590119 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd\": container with ID starting with 98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd not found: ID does not exist" containerID="98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.590170 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd"} err="failed to get container status \"98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd\": rpc error: code = NotFound desc = could not find container \"98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd\": container with ID starting with 98ca2535c39db41fb79d12be6af0f8c62f6a76c13f3d6a6d6bd5d2c590b768dd not found: ID does not exist" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.590195 4931 scope.go:117] "RemoveContainer" containerID="7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.594645 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-454st"] Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.610758 4931 scope.go:117] "RemoveContainer" containerID="7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753" Nov 25 22:03:42 crc kubenswrapper[4931]: E1125 22:03:42.611199 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753\": container with ID starting with 7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753 not found: ID does not exist" containerID="7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753" Nov 25 22:03:42 crc kubenswrapper[4931]: I1125 22:03:42.611228 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753"} err="failed to get container status \"7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753\": rpc error: code = NotFound desc = could not find container \"7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753\": container with ID starting with 7359bcd9b423a188126bc18a0a25755b07650995ed330979c3a275d9f2990753 not found: ID does not exist" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.417620 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5859867699-vd9ln"] Nov 25 22:03:43 crc kubenswrapper[4931]: E1125 22:03:43.418250 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74cbb6cf-4717-439c-b712-67718c154742" containerName="controller-manager" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.418270 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="74cbb6cf-4717-439c-b712-67718c154742" containerName="controller-manager" Nov 25 22:03:43 crc kubenswrapper[4931]: E1125 22:03:43.418288 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fada681c-6ade-49e4-b324-1985853bc61e" containerName="route-controller-manager" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.418299 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="fada681c-6ade-49e4-b324-1985853bc61e" containerName="route-controller-manager" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.418452 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="fada681c-6ade-49e4-b324-1985853bc61e" containerName="route-controller-manager" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.418478 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="74cbb6cf-4717-439c-b712-67718c154742" containerName="controller-manager" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.419017 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.422003 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.422261 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.422808 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.423202 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.423288 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.423731 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.425583 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5859867699-vd9ln"] Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.429678 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.452612 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9"] Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.453191 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.456992 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.457167 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.457278 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.457375 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.457593 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.457755 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.461901 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9"] Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.612902 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a025276a-f875-4851-8d79-7f4a18d070e7-config\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.613019 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82cc634b-1397-4459-ae50-4328b1429a32-serving-cert\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.613070 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlhkp\" (UniqueName: \"kubernetes.io/projected/82cc634b-1397-4459-ae50-4328b1429a32-kube-api-access-qlhkp\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.613155 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a025276a-f875-4851-8d79-7f4a18d070e7-proxy-ca-bundles\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.613188 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a025276a-f875-4851-8d79-7f4a18d070e7-client-ca\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.613234 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82cc634b-1397-4459-ae50-4328b1429a32-config\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.613262 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6b5l\" (UniqueName: \"kubernetes.io/projected/a025276a-f875-4851-8d79-7f4a18d070e7-kube-api-access-s6b5l\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.613317 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a025276a-f875-4851-8d79-7f4a18d070e7-serving-cert\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.613391 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82cc634b-1397-4459-ae50-4328b1429a32-client-ca\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.714046 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82cc634b-1397-4459-ae50-4328b1429a32-client-ca\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.714096 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a025276a-f875-4851-8d79-7f4a18d070e7-config\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.714123 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82cc634b-1397-4459-ae50-4328b1429a32-serving-cert\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.714144 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlhkp\" (UniqueName: \"kubernetes.io/projected/82cc634b-1397-4459-ae50-4328b1429a32-kube-api-access-qlhkp\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.714181 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a025276a-f875-4851-8d79-7f4a18d070e7-proxy-ca-bundles\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.714203 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a025276a-f875-4851-8d79-7f4a18d070e7-client-ca\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.714222 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82cc634b-1397-4459-ae50-4328b1429a32-config\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.714242 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6b5l\" (UniqueName: \"kubernetes.io/projected/a025276a-f875-4851-8d79-7f4a18d070e7-kube-api-access-s6b5l\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.714258 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a025276a-f875-4851-8d79-7f4a18d070e7-serving-cert\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.715379 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82cc634b-1397-4459-ae50-4328b1429a32-client-ca\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.715690 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a025276a-f875-4851-8d79-7f4a18d070e7-proxy-ca-bundles\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.715730 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a025276a-f875-4851-8d79-7f4a18d070e7-client-ca\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.716045 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82cc634b-1397-4459-ae50-4328b1429a32-config\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.716694 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a025276a-f875-4851-8d79-7f4a18d070e7-config\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.721277 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82cc634b-1397-4459-ae50-4328b1429a32-serving-cert\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.721378 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a025276a-f875-4851-8d79-7f4a18d070e7-serving-cert\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.733988 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlhkp\" (UniqueName: \"kubernetes.io/projected/82cc634b-1397-4459-ae50-4328b1429a32-kube-api-access-qlhkp\") pod \"route-controller-manager-6777cb45dc-zjml9\" (UID: \"82cc634b-1397-4459-ae50-4328b1429a32\") " pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.750076 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6b5l\" (UniqueName: \"kubernetes.io/projected/a025276a-f875-4851-8d79-7f4a18d070e7-kube-api-access-s6b5l\") pod \"controller-manager-5859867699-vd9ln\" (UID: \"a025276a-f875-4851-8d79-7f4a18d070e7\") " pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.792326 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:43 crc kubenswrapper[4931]: I1125 22:03:43.798640 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.015173 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9"] Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.082971 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5859867699-vd9ln"] Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.430003 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74cbb6cf-4717-439c-b712-67718c154742" path="/var/lib/kubelet/pods/74cbb6cf-4717-439c-b712-67718c154742/volumes" Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.431073 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fada681c-6ade-49e4-b324-1985853bc61e" path="/var/lib/kubelet/pods/fada681c-6ade-49e4-b324-1985853bc61e/volumes" Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.566350 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" event={"ID":"a025276a-f875-4851-8d79-7f4a18d070e7","Type":"ContainerStarted","Data":"18411a6511ba298fd77a763c7ab017315e9d76563ee05c28c1ded91103f6fca0"} Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.566416 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" event={"ID":"a025276a-f875-4851-8d79-7f4a18d070e7","Type":"ContainerStarted","Data":"0276972e376c3db2518e5c8f37c035d0acbc935d6887fe2915009440a59ee6de"} Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.566596 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.567839 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" event={"ID":"82cc634b-1397-4459-ae50-4328b1429a32","Type":"ContainerStarted","Data":"afd10f03540cbfddf20f9b6951a0b545fe1250681e0fa297e5bf3f821536b765"} Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.567886 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" event={"ID":"82cc634b-1397-4459-ae50-4328b1429a32","Type":"ContainerStarted","Data":"972eda728ca9ff8f1610c69ffa8db24168d480cadbfe7a71363acff2f78da062"} Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.568128 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.572793 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.574727 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.628368 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5859867699-vd9ln" podStartSLOduration=1.628350219 podStartE2EDuration="1.628350219s" podCreationTimestamp="2025-11-25 22:03:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:03:44.608454289 +0000 UTC m=+762.964966408" watchObservedRunningTime="2025-11-25 22:03:44.628350219 +0000 UTC m=+762.984862338" Nov 25 22:03:44 crc kubenswrapper[4931]: I1125 22:03:44.659424 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6777cb45dc-zjml9" podStartSLOduration=1.6593903380000001 podStartE2EDuration="1.659390338s" podCreationTimestamp="2025-11-25 22:03:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:03:44.656732464 +0000 UTC m=+763.013244583" watchObservedRunningTime="2025-11-25 22:03:44.659390338 +0000 UTC m=+763.015902447" Nov 25 22:03:46 crc kubenswrapper[4931]: I1125 22:03:46.833740 4931 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 22:03:57 crc kubenswrapper[4931]: I1125 22:03:57.647731 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.543186 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf"] Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.543955 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.546366 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-pmmv2" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.551334 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-wcxdv"] Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.553964 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.556792 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.557012 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.557429 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.570215 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf"] Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.646547 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qsmp8"] Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.647807 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.652299 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-wj9nf" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.652484 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.652540 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.652689 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.660480 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-xhr6z"] Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.661237 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.665123 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.674179 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-xhr6z"] Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.721529 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4l77\" (UniqueName: \"kubernetes.io/projected/fd84f23e-59df-4222-b73d-e291f7a18301-kube-api-access-n4l77\") pod \"frr-k8s-webhook-server-6998585d5-2wqcf\" (UID: \"fd84f23e-59df-4222-b73d-e291f7a18301\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.721809 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-reloader\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.721832 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b3601564-28ae-4877-8b80-33b5027bfcf5-frr-startup\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.721850 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-frr-conf\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.721868 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbj5x\" (UniqueName: \"kubernetes.io/projected/b3601564-28ae-4877-8b80-33b5027bfcf5-kube-api-access-jbj5x\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.722055 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-metrics\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.722127 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd84f23e-59df-4222-b73d-e291f7a18301-cert\") pod \"frr-k8s-webhook-server-6998585d5-2wqcf\" (UID: \"fd84f23e-59df-4222-b73d-e291f7a18301\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.722166 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3601564-28ae-4877-8b80-33b5027bfcf5-metrics-certs\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.722207 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-frr-sockets\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823361 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd84f23e-59df-4222-b73d-e291f7a18301-cert\") pod \"frr-k8s-webhook-server-6998585d5-2wqcf\" (UID: \"fd84f23e-59df-4222-b73d-e291f7a18301\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823407 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3601564-28ae-4877-8b80-33b5027bfcf5-metrics-certs\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823432 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c77cee34-eac7-4391-8fc0-f492acaa9004-metrics-certs\") pod \"controller-6c7b4b5f48-xhr6z\" (UID: \"c77cee34-eac7-4391-8fc0-f492acaa9004\") " pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823459 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlkcm\" (UniqueName: \"kubernetes.io/projected/c77cee34-eac7-4391-8fc0-f492acaa9004-kube-api-access-vlkcm\") pod \"controller-6c7b4b5f48-xhr6z\" (UID: \"c77cee34-eac7-4391-8fc0-f492acaa9004\") " pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823491 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hh7r\" (UniqueName: \"kubernetes.io/projected/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-kube-api-access-8hh7r\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823515 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c77cee34-eac7-4391-8fc0-f492acaa9004-cert\") pod \"controller-6c7b4b5f48-xhr6z\" (UID: \"c77cee34-eac7-4391-8fc0-f492acaa9004\") " pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823544 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-frr-sockets\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823570 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-metrics-certs\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823589 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4l77\" (UniqueName: \"kubernetes.io/projected/fd84f23e-59df-4222-b73d-e291f7a18301-kube-api-access-n4l77\") pod \"frr-k8s-webhook-server-6998585d5-2wqcf\" (UID: \"fd84f23e-59df-4222-b73d-e291f7a18301\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823607 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-reloader\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823624 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b3601564-28ae-4877-8b80-33b5027bfcf5-frr-startup\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823641 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-frr-conf\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823661 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbj5x\" (UniqueName: \"kubernetes.io/projected/b3601564-28ae-4877-8b80-33b5027bfcf5-kube-api-access-jbj5x\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823683 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-metallb-excludel2\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823701 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-metrics\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.823720 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-memberlist\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: E1125 22:03:58.823779 4931 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 25 22:03:58 crc kubenswrapper[4931]: E1125 22:03:58.823856 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3601564-28ae-4877-8b80-33b5027bfcf5-metrics-certs podName:b3601564-28ae-4877-8b80-33b5027bfcf5 nodeName:}" failed. No retries permitted until 2025-11-25 22:03:59.323830804 +0000 UTC m=+777.680342933 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3601564-28ae-4877-8b80-33b5027bfcf5-metrics-certs") pod "frr-k8s-wcxdv" (UID: "b3601564-28ae-4877-8b80-33b5027bfcf5") : secret "frr-k8s-certs-secret" not found Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.824119 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-frr-sockets\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.824329 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-frr-conf\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.824663 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-reloader\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.824789 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b3601564-28ae-4877-8b80-33b5027bfcf5-metrics\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.825079 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b3601564-28ae-4877-8b80-33b5027bfcf5-frr-startup\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.834807 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd84f23e-59df-4222-b73d-e291f7a18301-cert\") pod \"frr-k8s-webhook-server-6998585d5-2wqcf\" (UID: \"fd84f23e-59df-4222-b73d-e291f7a18301\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.841271 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbj5x\" (UniqueName: \"kubernetes.io/projected/b3601564-28ae-4877-8b80-33b5027bfcf5-kube-api-access-jbj5x\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.844622 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4l77\" (UniqueName: \"kubernetes.io/projected/fd84f23e-59df-4222-b73d-e291f7a18301-kube-api-access-n4l77\") pod \"frr-k8s-webhook-server-6998585d5-2wqcf\" (UID: \"fd84f23e-59df-4222-b73d-e291f7a18301\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.859400 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.925132 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c77cee34-eac7-4391-8fc0-f492acaa9004-metrics-certs\") pod \"controller-6c7b4b5f48-xhr6z\" (UID: \"c77cee34-eac7-4391-8fc0-f492acaa9004\") " pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.925174 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlkcm\" (UniqueName: \"kubernetes.io/projected/c77cee34-eac7-4391-8fc0-f492acaa9004-kube-api-access-vlkcm\") pod \"controller-6c7b4b5f48-xhr6z\" (UID: \"c77cee34-eac7-4391-8fc0-f492acaa9004\") " pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.925198 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hh7r\" (UniqueName: \"kubernetes.io/projected/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-kube-api-access-8hh7r\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.925219 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c77cee34-eac7-4391-8fc0-f492acaa9004-cert\") pod \"controller-6c7b4b5f48-xhr6z\" (UID: \"c77cee34-eac7-4391-8fc0-f492acaa9004\") " pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.925248 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-metrics-certs\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.925287 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-metallb-excludel2\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.925310 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-memberlist\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: E1125 22:03:58.925429 4931 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 22:03:58 crc kubenswrapper[4931]: E1125 22:03:58.925474 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-memberlist podName:15968cf8-d97e-4ba4-95c5-a3d82eb43e9c nodeName:}" failed. No retries permitted until 2025-11-25 22:03:59.425459362 +0000 UTC m=+777.781971481 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-memberlist") pod "speaker-qsmp8" (UID: "15968cf8-d97e-4ba4-95c5-a3d82eb43e9c") : secret "metallb-memberlist" not found Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.928240 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c77cee34-eac7-4391-8fc0-f492acaa9004-cert\") pod \"controller-6c7b4b5f48-xhr6z\" (UID: \"c77cee34-eac7-4391-8fc0-f492acaa9004\") " pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.928550 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-metrics-certs\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.929548 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c77cee34-eac7-4391-8fc0-f492acaa9004-metrics-certs\") pod \"controller-6c7b4b5f48-xhr6z\" (UID: \"c77cee34-eac7-4391-8fc0-f492acaa9004\") " pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.930046 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-metallb-excludel2\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.945179 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlkcm\" (UniqueName: \"kubernetes.io/projected/c77cee34-eac7-4391-8fc0-f492acaa9004-kube-api-access-vlkcm\") pod \"controller-6c7b4b5f48-xhr6z\" (UID: \"c77cee34-eac7-4391-8fc0-f492acaa9004\") " pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.957491 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hh7r\" (UniqueName: \"kubernetes.io/projected/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-kube-api-access-8hh7r\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:58 crc kubenswrapper[4931]: I1125 22:03:58.977752 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.274026 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf"] Nov 25 22:03:59 crc kubenswrapper[4931]: W1125 22:03:59.279167 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd84f23e_59df_4222_b73d_e291f7a18301.slice/crio-d817f3ebf9c5d9439416b4872ebf03f1c0fcd80469a2d005c2dde86f624b6d55 WatchSource:0}: Error finding container d817f3ebf9c5d9439416b4872ebf03f1c0fcd80469a2d005c2dde86f624b6d55: Status 404 returned error can't find the container with id d817f3ebf9c5d9439416b4872ebf03f1c0fcd80469a2d005c2dde86f624b6d55 Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.330497 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3601564-28ae-4877-8b80-33b5027bfcf5-metrics-certs\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.337772 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3601564-28ae-4877-8b80-33b5027bfcf5-metrics-certs\") pod \"frr-k8s-wcxdv\" (UID: \"b3601564-28ae-4877-8b80-33b5027bfcf5\") " pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.414549 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-xhr6z"] Nov 25 22:03:59 crc kubenswrapper[4931]: W1125 22:03:59.421350 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc77cee34_eac7_4391_8fc0_f492acaa9004.slice/crio-2aef1eec4c60f895f32be968ab62763b904053a5b765a7731e1b2892f1d09a23 WatchSource:0}: Error finding container 2aef1eec4c60f895f32be968ab62763b904053a5b765a7731e1b2892f1d09a23: Status 404 returned error can't find the container with id 2aef1eec4c60f895f32be968ab62763b904053a5b765a7731e1b2892f1d09a23 Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.431686 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-memberlist\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:03:59 crc kubenswrapper[4931]: E1125 22:03:59.431839 4931 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 22:03:59 crc kubenswrapper[4931]: E1125 22:03:59.432065 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-memberlist podName:15968cf8-d97e-4ba4-95c5-a3d82eb43e9c nodeName:}" failed. No retries permitted until 2025-11-25 22:04:00.432048602 +0000 UTC m=+778.788560721 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-memberlist") pod "speaker-qsmp8" (UID: "15968cf8-d97e-4ba4-95c5-a3d82eb43e9c") : secret "metallb-memberlist" not found Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.469506 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.668092 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerStarted","Data":"5a317af77ee4c2d242298d286d08ea101a6924d2cdb48de00bef36eacc2e9729"} Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.669784 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" event={"ID":"fd84f23e-59df-4222-b73d-e291f7a18301","Type":"ContainerStarted","Data":"d817f3ebf9c5d9439416b4872ebf03f1c0fcd80469a2d005c2dde86f624b6d55"} Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.671993 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-xhr6z" event={"ID":"c77cee34-eac7-4391-8fc0-f492acaa9004","Type":"ContainerStarted","Data":"c10670cdcdbbeb7fc358ba9046bf560d732432a315558ed1e5d587602f89399a"} Nov 25 22:03:59 crc kubenswrapper[4931]: I1125 22:03:59.672042 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-xhr6z" event={"ID":"c77cee34-eac7-4391-8fc0-f492acaa9004","Type":"ContainerStarted","Data":"2aef1eec4c60f895f32be968ab62763b904053a5b765a7731e1b2892f1d09a23"} Nov 25 22:04:00 crc kubenswrapper[4931]: I1125 22:04:00.443810 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-memberlist\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:04:00 crc kubenswrapper[4931]: I1125 22:04:00.451217 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/15968cf8-d97e-4ba4-95c5-a3d82eb43e9c-memberlist\") pod \"speaker-qsmp8\" (UID: \"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c\") " pod="metallb-system/speaker-qsmp8" Nov 25 22:04:00 crc kubenswrapper[4931]: I1125 22:04:00.466570 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qsmp8" Nov 25 22:04:00 crc kubenswrapper[4931]: W1125 22:04:00.489663 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15968cf8_d97e_4ba4_95c5_a3d82eb43e9c.slice/crio-a420e2e395e262bbffd3307ed8972e2d97897df1330c7c3f2e74c5879c09eb9d WatchSource:0}: Error finding container a420e2e395e262bbffd3307ed8972e2d97897df1330c7c3f2e74c5879c09eb9d: Status 404 returned error can't find the container with id a420e2e395e262bbffd3307ed8972e2d97897df1330c7c3f2e74c5879c09eb9d Nov 25 22:04:00 crc kubenswrapper[4931]: I1125 22:04:00.681658 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qsmp8" event={"ID":"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c","Type":"ContainerStarted","Data":"a420e2e395e262bbffd3307ed8972e2d97897df1330c7c3f2e74c5879c09eb9d"} Nov 25 22:04:01 crc kubenswrapper[4931]: I1125 22:04:01.688901 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qsmp8" event={"ID":"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c","Type":"ContainerStarted","Data":"a6479d07241d4a2b72e5d8700c0395d7bfc4f1f03a6639f5ecf72ed43cae4150"} Nov 25 22:04:03 crc kubenswrapper[4931]: I1125 22:04:03.704747 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qsmp8" event={"ID":"15968cf8-d97e-4ba4-95c5-a3d82eb43e9c","Type":"ContainerStarted","Data":"042eedbaae80c1e030f783af5b7372e244cbe173e65f6169af3a16014f6c0323"} Nov 25 22:04:03 crc kubenswrapper[4931]: I1125 22:04:03.705355 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qsmp8" Nov 25 22:04:03 crc kubenswrapper[4931]: I1125 22:04:03.707189 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-xhr6z" event={"ID":"c77cee34-eac7-4391-8fc0-f492acaa9004","Type":"ContainerStarted","Data":"69889c0b1f2766beec17f1c2df17681615d2d238183d4e6d99ccf539eb859079"} Nov 25 22:04:03 crc kubenswrapper[4931]: I1125 22:04:03.707459 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:04:03 crc kubenswrapper[4931]: I1125 22:04:03.724790 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qsmp8" podStartSLOduration=3.74564503 podStartE2EDuration="5.724770406s" podCreationTimestamp="2025-11-25 22:03:58 +0000 UTC" firstStartedPulling="2025-11-25 22:04:00.850720967 +0000 UTC m=+779.207233096" lastFinishedPulling="2025-11-25 22:04:02.829846353 +0000 UTC m=+781.186358472" observedRunningTime="2025-11-25 22:04:03.722745951 +0000 UTC m=+782.079258130" watchObservedRunningTime="2025-11-25 22:04:03.724770406 +0000 UTC m=+782.081282515" Nov 25 22:04:03 crc kubenswrapper[4931]: I1125 22:04:03.744077 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-xhr6z" podStartSLOduration=2.442322213 podStartE2EDuration="5.744053976s" podCreationTimestamp="2025-11-25 22:03:58 +0000 UTC" firstStartedPulling="2025-11-25 22:03:59.52476559 +0000 UTC m=+777.881277709" lastFinishedPulling="2025-11-25 22:04:02.826497353 +0000 UTC m=+781.183009472" observedRunningTime="2025-11-25 22:04:03.739223846 +0000 UTC m=+782.095735965" watchObservedRunningTime="2025-11-25 22:04:03.744053976 +0000 UTC m=+782.100566095" Nov 25 22:04:06 crc kubenswrapper[4931]: I1125 22:04:06.726321 4931 generic.go:334] "Generic (PLEG): container finished" podID="b3601564-28ae-4877-8b80-33b5027bfcf5" containerID="5f6755ec43e0304791c0e3c6f4a171ecdf1b1642d5823fb96f96f2abcc45163c" exitCode=0 Nov 25 22:04:06 crc kubenswrapper[4931]: I1125 22:04:06.726414 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerDied","Data":"5f6755ec43e0304791c0e3c6f4a171ecdf1b1642d5823fb96f96f2abcc45163c"} Nov 25 22:04:06 crc kubenswrapper[4931]: I1125 22:04:06.730643 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" event={"ID":"fd84f23e-59df-4222-b73d-e291f7a18301","Type":"ContainerStarted","Data":"e8860c8d588d1b982b30e58c7606f433e74dcd46e9834ac34bfc753f13a02061"} Nov 25 22:04:06 crc kubenswrapper[4931]: I1125 22:04:06.731051 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:04:06 crc kubenswrapper[4931]: I1125 22:04:06.786672 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" podStartSLOduration=1.606807271 podStartE2EDuration="8.786654437s" podCreationTimestamp="2025-11-25 22:03:58 +0000 UTC" firstStartedPulling="2025-11-25 22:03:59.281535146 +0000 UTC m=+777.638047265" lastFinishedPulling="2025-11-25 22:04:06.461382272 +0000 UTC m=+784.817894431" observedRunningTime="2025-11-25 22:04:06.783530472 +0000 UTC m=+785.140042611" watchObservedRunningTime="2025-11-25 22:04:06.786654437 +0000 UTC m=+785.143166566" Nov 25 22:04:07 crc kubenswrapper[4931]: I1125 22:04:07.740408 4931 generic.go:334] "Generic (PLEG): container finished" podID="b3601564-28ae-4877-8b80-33b5027bfcf5" containerID="cb13d551bb36ba69c8e60432156c73da7da58c72ebdcab884cbd670804650eea" exitCode=0 Nov 25 22:04:07 crc kubenswrapper[4931]: I1125 22:04:07.740529 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerDied","Data":"cb13d551bb36ba69c8e60432156c73da7da58c72ebdcab884cbd670804650eea"} Nov 25 22:04:08 crc kubenswrapper[4931]: I1125 22:04:08.751179 4931 generic.go:334] "Generic (PLEG): container finished" podID="b3601564-28ae-4877-8b80-33b5027bfcf5" containerID="2bc2644281ee81dd857ab6fe205a733c063a25473c6a097d014755b7a0d6e077" exitCode=0 Nov 25 22:04:08 crc kubenswrapper[4931]: I1125 22:04:08.751291 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerDied","Data":"2bc2644281ee81dd857ab6fe205a733c063a25473c6a097d014755b7a0d6e077"} Nov 25 22:04:08 crc kubenswrapper[4931]: I1125 22:04:08.843528 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:04:08 crc kubenswrapper[4931]: I1125 22:04:08.843617 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:04:09 crc kubenswrapper[4931]: I1125 22:04:09.764145 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerStarted","Data":"b48c3e96e23549cb29587896764475ee3689b07b4c6e820a59ca1839ec4bf016"} Nov 25 22:04:09 crc kubenswrapper[4931]: I1125 22:04:09.764201 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerStarted","Data":"5d10cf6f74b55b27ae62e574f84e513f2f285793e836d103b8bd5e3d068630e2"} Nov 25 22:04:09 crc kubenswrapper[4931]: I1125 22:04:09.764215 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerStarted","Data":"5eca5c4c9c07b7e6bb681ccc15d358efede5b80406e62f15ddd1fd6aa5dc30b5"} Nov 25 22:04:09 crc kubenswrapper[4931]: I1125 22:04:09.764227 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerStarted","Data":"c32fc6e3424082cacb652dd5d208a4ef9c680475a250884f2b716c597a039b64"} Nov 25 22:04:09 crc kubenswrapper[4931]: I1125 22:04:09.764238 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerStarted","Data":"7914c6b2c95a64d78706231bca46fb8fac07ab10ec86379736155173115a3f7b"} Nov 25 22:04:10 crc kubenswrapper[4931]: I1125 22:04:10.472154 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qsmp8" Nov 25 22:04:10 crc kubenswrapper[4931]: I1125 22:04:10.801697 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wcxdv" event={"ID":"b3601564-28ae-4877-8b80-33b5027bfcf5","Type":"ContainerStarted","Data":"6b05e4730d7c1301f89b4ee9324036adfa00f0ae4e4c260572d368afcbac9e3c"} Nov 25 22:04:10 crc kubenswrapper[4931]: I1125 22:04:10.802049 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:04:10 crc kubenswrapper[4931]: I1125 22:04:10.856723 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-wcxdv" podStartSLOduration=5.968998987 podStartE2EDuration="12.856694331s" podCreationTimestamp="2025-11-25 22:03:58 +0000 UTC" firstStartedPulling="2025-11-25 22:03:59.55332728 +0000 UTC m=+777.909839419" lastFinishedPulling="2025-11-25 22:04:06.441022634 +0000 UTC m=+784.797534763" observedRunningTime="2025-11-25 22:04:10.85256761 +0000 UTC m=+789.209079729" watchObservedRunningTime="2025-11-25 22:04:10.856694331 +0000 UTC m=+789.213206450" Nov 25 22:04:14 crc kubenswrapper[4931]: I1125 22:04:14.470328 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:04:14 crc kubenswrapper[4931]: I1125 22:04:14.506371 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:04:17 crc kubenswrapper[4931]: I1125 22:04:17.235349 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-59qxk"] Nov 25 22:04:17 crc kubenswrapper[4931]: I1125 22:04:17.236130 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-59qxk" Nov 25 22:04:17 crc kubenswrapper[4931]: I1125 22:04:17.239029 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-lfhjt" Nov 25 22:04:17 crc kubenswrapper[4931]: I1125 22:04:17.239958 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 22:04:17 crc kubenswrapper[4931]: I1125 22:04:17.241784 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 22:04:17 crc kubenswrapper[4931]: I1125 22:04:17.263376 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-59qxk"] Nov 25 22:04:17 crc kubenswrapper[4931]: I1125 22:04:17.419450 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6s4g\" (UniqueName: \"kubernetes.io/projected/9876b633-f436-4211-b88b-d2e58aafb00d-kube-api-access-m6s4g\") pod \"mariadb-operator-index-59qxk\" (UID: \"9876b633-f436-4211-b88b-d2e58aafb00d\") " pod="openstack-operators/mariadb-operator-index-59qxk" Nov 25 22:04:17 crc kubenswrapper[4931]: I1125 22:04:17.521544 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6s4g\" (UniqueName: \"kubernetes.io/projected/9876b633-f436-4211-b88b-d2e58aafb00d-kube-api-access-m6s4g\") pod \"mariadb-operator-index-59qxk\" (UID: \"9876b633-f436-4211-b88b-d2e58aafb00d\") " pod="openstack-operators/mariadb-operator-index-59qxk" Nov 25 22:04:18 crc kubenswrapper[4931]: I1125 22:04:18.285219 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6s4g\" (UniqueName: \"kubernetes.io/projected/9876b633-f436-4211-b88b-d2e58aafb00d-kube-api-access-m6s4g\") pod \"mariadb-operator-index-59qxk\" (UID: \"9876b633-f436-4211-b88b-d2e58aafb00d\") " pod="openstack-operators/mariadb-operator-index-59qxk" Nov 25 22:04:18 crc kubenswrapper[4931]: I1125 22:04:18.452692 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-59qxk" Nov 25 22:04:18 crc kubenswrapper[4931]: I1125 22:04:18.870565 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-2wqcf" Nov 25 22:04:18 crc kubenswrapper[4931]: I1125 22:04:18.958837 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-59qxk"] Nov 25 22:04:18 crc kubenswrapper[4931]: W1125 22:04:18.966076 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9876b633_f436_4211_b88b_d2e58aafb00d.slice/crio-3dd8607452eb75312918e80121974e709a254de36ddf9a4e83f25a4eb725e3aa WatchSource:0}: Error finding container 3dd8607452eb75312918e80121974e709a254de36ddf9a4e83f25a4eb725e3aa: Status 404 returned error can't find the container with id 3dd8607452eb75312918e80121974e709a254de36ddf9a4e83f25a4eb725e3aa Nov 25 22:04:18 crc kubenswrapper[4931]: I1125 22:04:18.981380 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-xhr6z" Nov 25 22:04:19 crc kubenswrapper[4931]: I1125 22:04:19.273179 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-59qxk" event={"ID":"9876b633-f436-4211-b88b-d2e58aafb00d","Type":"ContainerStarted","Data":"3dd8607452eb75312918e80121974e709a254de36ddf9a4e83f25a4eb725e3aa"} Nov 25 22:04:19 crc kubenswrapper[4931]: I1125 22:04:19.475196 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wcxdv" Nov 25 22:04:20 crc kubenswrapper[4931]: I1125 22:04:20.280271 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-59qxk" event={"ID":"9876b633-f436-4211-b88b-d2e58aafb00d","Type":"ContainerStarted","Data":"5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d"} Nov 25 22:04:20 crc kubenswrapper[4931]: I1125 22:04:20.299595 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-59qxk" podStartSLOduration=2.281718587 podStartE2EDuration="3.299572472s" podCreationTimestamp="2025-11-25 22:04:17 +0000 UTC" firstStartedPulling="2025-11-25 22:04:18.967378697 +0000 UTC m=+797.323890816" lastFinishedPulling="2025-11-25 22:04:19.985232542 +0000 UTC m=+798.341744701" observedRunningTime="2025-11-25 22:04:20.293736025 +0000 UTC m=+798.650248154" watchObservedRunningTime="2025-11-25 22:04:20.299572472 +0000 UTC m=+798.656084601" Nov 25 22:04:20 crc kubenswrapper[4931]: I1125 22:04:20.791491 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-59qxk"] Nov 25 22:04:21 crc kubenswrapper[4931]: I1125 22:04:21.397198 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-znxs7"] Nov 25 22:04:21 crc kubenswrapper[4931]: I1125 22:04:21.398515 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-znxs7" Nov 25 22:04:21 crc kubenswrapper[4931]: I1125 22:04:21.419101 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-znxs7"] Nov 25 22:04:21 crc kubenswrapper[4931]: I1125 22:04:21.485839 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2ssc\" (UniqueName: \"kubernetes.io/projected/413e27a3-fd17-4743-9ac4-6767858ea041-kube-api-access-w2ssc\") pod \"mariadb-operator-index-znxs7\" (UID: \"413e27a3-fd17-4743-9ac4-6767858ea041\") " pod="openstack-operators/mariadb-operator-index-znxs7" Nov 25 22:04:21 crc kubenswrapper[4931]: I1125 22:04:21.586924 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2ssc\" (UniqueName: \"kubernetes.io/projected/413e27a3-fd17-4743-9ac4-6767858ea041-kube-api-access-w2ssc\") pod \"mariadb-operator-index-znxs7\" (UID: \"413e27a3-fd17-4743-9ac4-6767858ea041\") " pod="openstack-operators/mariadb-operator-index-znxs7" Nov 25 22:04:21 crc kubenswrapper[4931]: I1125 22:04:21.621683 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2ssc\" (UniqueName: \"kubernetes.io/projected/413e27a3-fd17-4743-9ac4-6767858ea041-kube-api-access-w2ssc\") pod \"mariadb-operator-index-znxs7\" (UID: \"413e27a3-fd17-4743-9ac4-6767858ea041\") " pod="openstack-operators/mariadb-operator-index-znxs7" Nov 25 22:04:21 crc kubenswrapper[4931]: I1125 22:04:21.730705 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-znxs7" Nov 25 22:04:22 crc kubenswrapper[4931]: I1125 22:04:22.246328 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-znxs7"] Nov 25 22:04:22 crc kubenswrapper[4931]: W1125 22:04:22.258507 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod413e27a3_fd17_4743_9ac4_6767858ea041.slice/crio-5a9e388165fa82c248034b933e97d99bb2a54ea6490f29d397f698edc1f4595b WatchSource:0}: Error finding container 5a9e388165fa82c248034b933e97d99bb2a54ea6490f29d397f698edc1f4595b: Status 404 returned error can't find the container with id 5a9e388165fa82c248034b933e97d99bb2a54ea6490f29d397f698edc1f4595b Nov 25 22:04:22 crc kubenswrapper[4931]: I1125 22:04:22.297428 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-znxs7" event={"ID":"413e27a3-fd17-4743-9ac4-6767858ea041","Type":"ContainerStarted","Data":"5a9e388165fa82c248034b933e97d99bb2a54ea6490f29d397f698edc1f4595b"} Nov 25 22:04:22 crc kubenswrapper[4931]: I1125 22:04:22.297694 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-59qxk" podUID="9876b633-f436-4211-b88b-d2e58aafb00d" containerName="registry-server" containerID="cri-o://5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d" gracePeriod=2 Nov 25 22:04:22 crc kubenswrapper[4931]: I1125 22:04:22.889000 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-59qxk" Nov 25 22:04:22 crc kubenswrapper[4931]: I1125 22:04:22.912291 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6s4g\" (UniqueName: \"kubernetes.io/projected/9876b633-f436-4211-b88b-d2e58aafb00d-kube-api-access-m6s4g\") pod \"9876b633-f436-4211-b88b-d2e58aafb00d\" (UID: \"9876b633-f436-4211-b88b-d2e58aafb00d\") " Nov 25 22:04:22 crc kubenswrapper[4931]: I1125 22:04:22.921493 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9876b633-f436-4211-b88b-d2e58aafb00d-kube-api-access-m6s4g" (OuterVolumeSpecName: "kube-api-access-m6s4g") pod "9876b633-f436-4211-b88b-d2e58aafb00d" (UID: "9876b633-f436-4211-b88b-d2e58aafb00d"). InnerVolumeSpecName "kube-api-access-m6s4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.013984 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6s4g\" (UniqueName: \"kubernetes.io/projected/9876b633-f436-4211-b88b-d2e58aafb00d-kube-api-access-m6s4g\") on node \"crc\" DevicePath \"\"" Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.304601 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-znxs7" event={"ID":"413e27a3-fd17-4743-9ac4-6767858ea041","Type":"ContainerStarted","Data":"5fac6841d4a323afb93d8f903d3be8e8ce73a96111230f32c3cb3137d9a72731"} Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.307583 4931 generic.go:334] "Generic (PLEG): container finished" podID="9876b633-f436-4211-b88b-d2e58aafb00d" containerID="5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d" exitCode=0 Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.307630 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-59qxk" Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.307631 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-59qxk" event={"ID":"9876b633-f436-4211-b88b-d2e58aafb00d","Type":"ContainerDied","Data":"5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d"} Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.307796 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-59qxk" event={"ID":"9876b633-f436-4211-b88b-d2e58aafb00d","Type":"ContainerDied","Data":"3dd8607452eb75312918e80121974e709a254de36ddf9a4e83f25a4eb725e3aa"} Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.307829 4931 scope.go:117] "RemoveContainer" containerID="5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d" Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.326325 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-znxs7" podStartSLOduration=1.836833068 podStartE2EDuration="2.326305726s" podCreationTimestamp="2025-11-25 22:04:21 +0000 UTC" firstStartedPulling="2025-11-25 22:04:22.265793071 +0000 UTC m=+800.622305230" lastFinishedPulling="2025-11-25 22:04:22.755265759 +0000 UTC m=+801.111777888" observedRunningTime="2025-11-25 22:04:23.321614379 +0000 UTC m=+801.678126518" watchObservedRunningTime="2025-11-25 22:04:23.326305726 +0000 UTC m=+801.682817855" Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.344375 4931 scope.go:117] "RemoveContainer" containerID="5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d" Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.346867 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-59qxk"] Nov 25 22:04:23 crc kubenswrapper[4931]: E1125 22:04:23.348340 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d\": container with ID starting with 5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d not found: ID does not exist" containerID="5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d" Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.348381 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d"} err="failed to get container status \"5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d\": rpc error: code = NotFound desc = could not find container \"5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d\": container with ID starting with 5ba5597ee0dbb31781629818028bcf40310ef0da5df8692d6fc6675aaaeed75d not found: ID does not exist" Nov 25 22:04:23 crc kubenswrapper[4931]: I1125 22:04:23.349733 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-59qxk"] Nov 25 22:04:23 crc kubenswrapper[4931]: E1125 22:04:23.373081 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9876b633_f436_4211_b88b_d2e58aafb00d.slice\": RecentStats: unable to find data in memory cache]" Nov 25 22:04:24 crc kubenswrapper[4931]: I1125 22:04:24.438158 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9876b633-f436-4211-b88b-d2e58aafb00d" path="/var/lib/kubelet/pods/9876b633-f436-4211-b88b-d2e58aafb00d/volumes" Nov 25 22:04:31 crc kubenswrapper[4931]: I1125 22:04:31.731602 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-znxs7" Nov 25 22:04:31 crc kubenswrapper[4931]: I1125 22:04:31.732292 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-znxs7" Nov 25 22:04:31 crc kubenswrapper[4931]: I1125 22:04:31.779539 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-znxs7" Nov 25 22:04:32 crc kubenswrapper[4931]: I1125 22:04:32.412470 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-znxs7" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.605812 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5"] Nov 25 22:04:38 crc kubenswrapper[4931]: E1125 22:04:38.606943 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9876b633-f436-4211-b88b-d2e58aafb00d" containerName="registry-server" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.606970 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9876b633-f436-4211-b88b-d2e58aafb00d" containerName="registry-server" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.607162 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9876b633-f436-4211-b88b-d2e58aafb00d" containerName="registry-server" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.608544 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.610891 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrsb7" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.611848 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5"] Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.744086 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.744297 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7st7\" (UniqueName: \"kubernetes.io/projected/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-kube-api-access-k7st7\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.744440 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.844228 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.844369 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.846212 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7st7\" (UniqueName: \"kubernetes.io/projected/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-kube-api-access-k7st7\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.846286 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.846387 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.846819 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.847761 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.869849 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7st7\" (UniqueName: \"kubernetes.io/projected/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-kube-api-access-k7st7\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:38 crc kubenswrapper[4931]: I1125 22:04:38.934328 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:39 crc kubenswrapper[4931]: I1125 22:04:39.379050 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5"] Nov 25 22:04:39 crc kubenswrapper[4931]: I1125 22:04:39.456557 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" event={"ID":"b355cbaf-1e51-49b8-a0a7-5591d6d62f25","Type":"ContainerStarted","Data":"630e644716673f6a05b32f685031af1c08fa20b1a26039bbbdcb6aacbc1d7624"} Nov 25 22:04:40 crc kubenswrapper[4931]: I1125 22:04:40.469765 4931 generic.go:334] "Generic (PLEG): container finished" podID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerID="6678304e8d2a7eabd8d4a45b31fc0ddc5d30bdf07d19fc609444f60a7abcd66e" exitCode=0 Nov 25 22:04:40 crc kubenswrapper[4931]: I1125 22:04:40.469843 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" event={"ID":"b355cbaf-1e51-49b8-a0a7-5591d6d62f25","Type":"ContainerDied","Data":"6678304e8d2a7eabd8d4a45b31fc0ddc5d30bdf07d19fc609444f60a7abcd66e"} Nov 25 22:04:41 crc kubenswrapper[4931]: I1125 22:04:41.485910 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" event={"ID":"b355cbaf-1e51-49b8-a0a7-5591d6d62f25","Type":"ContainerStarted","Data":"190e63cd5fe6fac6a24ea9b7d6a1e18b78257d24a16c4041f3313e87629dc9f8"} Nov 25 22:04:42 crc kubenswrapper[4931]: I1125 22:04:42.494100 4931 generic.go:334] "Generic (PLEG): container finished" podID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerID="190e63cd5fe6fac6a24ea9b7d6a1e18b78257d24a16c4041f3313e87629dc9f8" exitCode=0 Nov 25 22:04:42 crc kubenswrapper[4931]: I1125 22:04:42.494149 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" event={"ID":"b355cbaf-1e51-49b8-a0a7-5591d6d62f25","Type":"ContainerDied","Data":"190e63cd5fe6fac6a24ea9b7d6a1e18b78257d24a16c4041f3313e87629dc9f8"} Nov 25 22:04:43 crc kubenswrapper[4931]: I1125 22:04:43.504704 4931 generic.go:334] "Generic (PLEG): container finished" podID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerID="2259b4be9464a4e91a5be1d584a48131015372aa22e7435681433ca2f84de70b" exitCode=0 Nov 25 22:04:43 crc kubenswrapper[4931]: I1125 22:04:43.504749 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" event={"ID":"b355cbaf-1e51-49b8-a0a7-5591d6d62f25","Type":"ContainerDied","Data":"2259b4be9464a4e91a5be1d584a48131015372aa22e7435681433ca2f84de70b"} Nov 25 22:04:44 crc kubenswrapper[4931]: I1125 22:04:44.843408 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:44 crc kubenswrapper[4931]: I1125 22:04:44.931153 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-util\") pod \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " Nov 25 22:04:44 crc kubenswrapper[4931]: I1125 22:04:44.931260 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7st7\" (UniqueName: \"kubernetes.io/projected/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-kube-api-access-k7st7\") pod \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " Nov 25 22:04:44 crc kubenswrapper[4931]: I1125 22:04:44.931311 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-bundle\") pod \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\" (UID: \"b355cbaf-1e51-49b8-a0a7-5591d6d62f25\") " Nov 25 22:04:44 crc kubenswrapper[4931]: I1125 22:04:44.932311 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-bundle" (OuterVolumeSpecName: "bundle") pod "b355cbaf-1e51-49b8-a0a7-5591d6d62f25" (UID: "b355cbaf-1e51-49b8-a0a7-5591d6d62f25"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:04:44 crc kubenswrapper[4931]: I1125 22:04:44.948173 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-kube-api-access-k7st7" (OuterVolumeSpecName: "kube-api-access-k7st7") pod "b355cbaf-1e51-49b8-a0a7-5591d6d62f25" (UID: "b355cbaf-1e51-49b8-a0a7-5591d6d62f25"). InnerVolumeSpecName "kube-api-access-k7st7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:04:44 crc kubenswrapper[4931]: I1125 22:04:44.952773 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-util" (OuterVolumeSpecName: "util") pod "b355cbaf-1e51-49b8-a0a7-5591d6d62f25" (UID: "b355cbaf-1e51-49b8-a0a7-5591d6d62f25"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:04:45 crc kubenswrapper[4931]: I1125 22:04:45.033253 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-util\") on node \"crc\" DevicePath \"\"" Nov 25 22:04:45 crc kubenswrapper[4931]: I1125 22:04:45.033290 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7st7\" (UniqueName: \"kubernetes.io/projected/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-kube-api-access-k7st7\") on node \"crc\" DevicePath \"\"" Nov 25 22:04:45 crc kubenswrapper[4931]: I1125 22:04:45.033307 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b355cbaf-1e51-49b8-a0a7-5591d6d62f25-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:04:45 crc kubenswrapper[4931]: I1125 22:04:45.524446 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" event={"ID":"b355cbaf-1e51-49b8-a0a7-5591d6d62f25","Type":"ContainerDied","Data":"630e644716673f6a05b32f685031af1c08fa20b1a26039bbbdcb6aacbc1d7624"} Nov 25 22:04:45 crc kubenswrapper[4931]: I1125 22:04:45.524487 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="630e644716673f6a05b32f685031af1c08fa20b1a26039bbbdcb6aacbc1d7624" Nov 25 22:04:45 crc kubenswrapper[4931]: I1125 22:04:45.524561 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.798385 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv"] Nov 25 22:04:51 crc kubenswrapper[4931]: E1125 22:04:51.799257 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerName="util" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.799276 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerName="util" Nov 25 22:04:51 crc kubenswrapper[4931]: E1125 22:04:51.799291 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerName="pull" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.799298 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerName="pull" Nov 25 22:04:51 crc kubenswrapper[4931]: E1125 22:04:51.799315 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerName="extract" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.799321 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerName="extract" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.799415 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b355cbaf-1e51-49b8-a0a7-5591d6d62f25" containerName="extract" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.799811 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.801257 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-prgr5" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.802274 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.802515 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.847672 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv"] Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.924674 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/952695c7-9eff-4e33-8777-6a5ea50a2b94-webhook-cert\") pod \"mariadb-operator-controller-manager-7b69997d9b-sz4qv\" (UID: \"952695c7-9eff-4e33-8777-6a5ea50a2b94\") " pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.924737 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/952695c7-9eff-4e33-8777-6a5ea50a2b94-apiservice-cert\") pod \"mariadb-operator-controller-manager-7b69997d9b-sz4qv\" (UID: \"952695c7-9eff-4e33-8777-6a5ea50a2b94\") " pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:51 crc kubenswrapper[4931]: I1125 22:04:51.924801 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2md6\" (UniqueName: \"kubernetes.io/projected/952695c7-9eff-4e33-8777-6a5ea50a2b94-kube-api-access-l2md6\") pod \"mariadb-operator-controller-manager-7b69997d9b-sz4qv\" (UID: \"952695c7-9eff-4e33-8777-6a5ea50a2b94\") " pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:52 crc kubenswrapper[4931]: I1125 22:04:52.025647 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2md6\" (UniqueName: \"kubernetes.io/projected/952695c7-9eff-4e33-8777-6a5ea50a2b94-kube-api-access-l2md6\") pod \"mariadb-operator-controller-manager-7b69997d9b-sz4qv\" (UID: \"952695c7-9eff-4e33-8777-6a5ea50a2b94\") " pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:52 crc kubenswrapper[4931]: I1125 22:04:52.025789 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/952695c7-9eff-4e33-8777-6a5ea50a2b94-webhook-cert\") pod \"mariadb-operator-controller-manager-7b69997d9b-sz4qv\" (UID: \"952695c7-9eff-4e33-8777-6a5ea50a2b94\") " pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:52 crc kubenswrapper[4931]: I1125 22:04:52.025844 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/952695c7-9eff-4e33-8777-6a5ea50a2b94-apiservice-cert\") pod \"mariadb-operator-controller-manager-7b69997d9b-sz4qv\" (UID: \"952695c7-9eff-4e33-8777-6a5ea50a2b94\") " pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:52 crc kubenswrapper[4931]: I1125 22:04:52.034120 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/952695c7-9eff-4e33-8777-6a5ea50a2b94-apiservice-cert\") pod \"mariadb-operator-controller-manager-7b69997d9b-sz4qv\" (UID: \"952695c7-9eff-4e33-8777-6a5ea50a2b94\") " pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:52 crc kubenswrapper[4931]: I1125 22:04:52.039733 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/952695c7-9eff-4e33-8777-6a5ea50a2b94-webhook-cert\") pod \"mariadb-operator-controller-manager-7b69997d9b-sz4qv\" (UID: \"952695c7-9eff-4e33-8777-6a5ea50a2b94\") " pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:52 crc kubenswrapper[4931]: I1125 22:04:52.042800 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2md6\" (UniqueName: \"kubernetes.io/projected/952695c7-9eff-4e33-8777-6a5ea50a2b94-kube-api-access-l2md6\") pod \"mariadb-operator-controller-manager-7b69997d9b-sz4qv\" (UID: \"952695c7-9eff-4e33-8777-6a5ea50a2b94\") " pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:52 crc kubenswrapper[4931]: I1125 22:04:52.114588 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:52 crc kubenswrapper[4931]: I1125 22:04:52.550937 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv"] Nov 25 22:04:52 crc kubenswrapper[4931]: I1125 22:04:52.565727 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" event={"ID":"952695c7-9eff-4e33-8777-6a5ea50a2b94","Type":"ContainerStarted","Data":"c0bf08f4ec46f93152091da2ee9b5f70c435bddda6dc45cc73c707ea9cffae44"} Nov 25 22:04:56 crc kubenswrapper[4931]: I1125 22:04:56.600235 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" event={"ID":"952695c7-9eff-4e33-8777-6a5ea50a2b94","Type":"ContainerStarted","Data":"e124225a0871c7f1bdd8ce73d78764c990f06ae1ed18feb2546f4f0cfbd20973"} Nov 25 22:04:56 crc kubenswrapper[4931]: I1125 22:04:56.601017 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:04:56 crc kubenswrapper[4931]: I1125 22:04:56.624707 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" podStartSLOduration=2.07494598 podStartE2EDuration="5.624682006s" podCreationTimestamp="2025-11-25 22:04:51 +0000 UTC" firstStartedPulling="2025-11-25 22:04:52.558448604 +0000 UTC m=+830.914960723" lastFinishedPulling="2025-11-25 22:04:56.10818464 +0000 UTC m=+834.464696749" observedRunningTime="2025-11-25 22:04:56.620766291 +0000 UTC m=+834.977278420" watchObservedRunningTime="2025-11-25 22:04:56.624682006 +0000 UTC m=+834.981194145" Nov 25 22:05:02 crc kubenswrapper[4931]: I1125 22:05:02.119441 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:05:08 crc kubenswrapper[4931]: I1125 22:05:08.843427 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:05:08 crc kubenswrapper[4931]: I1125 22:05:08.843866 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:05:08 crc kubenswrapper[4931]: I1125 22:05:08.843931 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 22:05:08 crc kubenswrapper[4931]: I1125 22:05:08.844544 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e502eeb6e6462755666e0a9f82f8ea7195e2305ad1bd997e1c4a8dc5d91dd65a"} pod="openshift-machine-config-operator/machine-config-daemon-djjhg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 22:05:08 crc kubenswrapper[4931]: I1125 22:05:08.844609 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" containerID="cri-o://e502eeb6e6462755666e0a9f82f8ea7195e2305ad1bd997e1c4a8dc5d91dd65a" gracePeriod=600 Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.683666 4931 generic.go:334] "Generic (PLEG): container finished" podID="e22448de-5122-43be-853a-081347c13cf2" containerID="e502eeb6e6462755666e0a9f82f8ea7195e2305ad1bd997e1c4a8dc5d91dd65a" exitCode=0 Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.683752 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerDied","Data":"e502eeb6e6462755666e0a9f82f8ea7195e2305ad1bd997e1c4a8dc5d91dd65a"} Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.684149 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"f284a26c69ccc770f75121f3dbe719003137727be2b936d7ebbd69a2ec135257"} Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.684189 4931 scope.go:117] "RemoveContainer" containerID="e3f51442d6d51ca472d5b7468f6a60a5a87ed110feeee3ae340f88e811a0b5eb" Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.710711 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nqrm9"] Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.712722 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.724672 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nqrm9"] Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.879985 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2qqm\" (UniqueName: \"kubernetes.io/projected/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-kube-api-access-f2qqm\") pod \"redhat-operators-nqrm9\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.881042 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-catalog-content\") pod \"redhat-operators-nqrm9\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.881102 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-utilities\") pod \"redhat-operators-nqrm9\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.982800 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-utilities\") pod \"redhat-operators-nqrm9\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.982979 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2qqm\" (UniqueName: \"kubernetes.io/projected/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-kube-api-access-f2qqm\") pod \"redhat-operators-nqrm9\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.983021 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-catalog-content\") pod \"redhat-operators-nqrm9\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.984309 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-utilities\") pod \"redhat-operators-nqrm9\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:09 crc kubenswrapper[4931]: I1125 22:05:09.984363 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-catalog-content\") pod \"redhat-operators-nqrm9\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.012568 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2qqm\" (UniqueName: \"kubernetes.io/projected/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-kube-api-access-f2qqm\") pod \"redhat-operators-nqrm9\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.043122 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.308164 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nqrm9"] Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.690431 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-kv5fj"] Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.691642 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-kv5fj" Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.694105 4931 generic.go:334] "Generic (PLEG): container finished" podID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerID="05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0" exitCode=0 Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.694158 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqrm9" event={"ID":"dd248e3a-0fd2-4616-8b2f-c9e5671650ce","Type":"ContainerDied","Data":"05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0"} Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.694190 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqrm9" event={"ID":"dd248e3a-0fd2-4616-8b2f-c9e5671650ce","Type":"ContainerStarted","Data":"f8bed3e9c13b8d5809a16339b58e1fcaa312280d6edcb26232208817cefab16b"} Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.694829 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-2jtxv" Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.711653 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-kv5fj"] Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.792357 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ptp9\" (UniqueName: \"kubernetes.io/projected/3f1ba234-ef51-46fd-81ba-16db04e0c689-kube-api-access-4ptp9\") pod \"infra-operator-index-kv5fj\" (UID: \"3f1ba234-ef51-46fd-81ba-16db04e0c689\") " pod="openstack-operators/infra-operator-index-kv5fj" Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.893250 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ptp9\" (UniqueName: \"kubernetes.io/projected/3f1ba234-ef51-46fd-81ba-16db04e0c689-kube-api-access-4ptp9\") pod \"infra-operator-index-kv5fj\" (UID: \"3f1ba234-ef51-46fd-81ba-16db04e0c689\") " pod="openstack-operators/infra-operator-index-kv5fj" Nov 25 22:05:10 crc kubenswrapper[4931]: I1125 22:05:10.922732 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ptp9\" (UniqueName: \"kubernetes.io/projected/3f1ba234-ef51-46fd-81ba-16db04e0c689-kube-api-access-4ptp9\") pod \"infra-operator-index-kv5fj\" (UID: \"3f1ba234-ef51-46fd-81ba-16db04e0c689\") " pod="openstack-operators/infra-operator-index-kv5fj" Nov 25 22:05:11 crc kubenswrapper[4931]: I1125 22:05:11.023790 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-kv5fj" Nov 25 22:05:11 crc kubenswrapper[4931]: I1125 22:05:11.455262 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-kv5fj"] Nov 25 22:05:11 crc kubenswrapper[4931]: W1125 22:05:11.463570 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f1ba234_ef51_46fd_81ba_16db04e0c689.slice/crio-8170bb7e63502ba30db8efcfb7dc59a0a0e3abc6eb843f5eb03f259b412978f4 WatchSource:0}: Error finding container 8170bb7e63502ba30db8efcfb7dc59a0a0e3abc6eb843f5eb03f259b412978f4: Status 404 returned error can't find the container with id 8170bb7e63502ba30db8efcfb7dc59a0a0e3abc6eb843f5eb03f259b412978f4 Nov 25 22:05:11 crc kubenswrapper[4931]: I1125 22:05:11.703484 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqrm9" event={"ID":"dd248e3a-0fd2-4616-8b2f-c9e5671650ce","Type":"ContainerStarted","Data":"1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0"} Nov 25 22:05:11 crc kubenswrapper[4931]: I1125 22:05:11.704753 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-kv5fj" event={"ID":"3f1ba234-ef51-46fd-81ba-16db04e0c689","Type":"ContainerStarted","Data":"8170bb7e63502ba30db8efcfb7dc59a0a0e3abc6eb843f5eb03f259b412978f4"} Nov 25 22:05:12 crc kubenswrapper[4931]: I1125 22:05:12.720990 4931 generic.go:334] "Generic (PLEG): container finished" podID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerID="1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0" exitCode=0 Nov 25 22:05:12 crc kubenswrapper[4931]: I1125 22:05:12.721054 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqrm9" event={"ID":"dd248e3a-0fd2-4616-8b2f-c9e5671650ce","Type":"ContainerDied","Data":"1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0"} Nov 25 22:05:13 crc kubenswrapper[4931]: I1125 22:05:13.731522 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqrm9" event={"ID":"dd248e3a-0fd2-4616-8b2f-c9e5671650ce","Type":"ContainerStarted","Data":"2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3"} Nov 25 22:05:13 crc kubenswrapper[4931]: I1125 22:05:13.733571 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-kv5fj" event={"ID":"3f1ba234-ef51-46fd-81ba-16db04e0c689","Type":"ContainerStarted","Data":"96c779cfc9ec8bc247664ad909e8cfb76541307d61872255fdb95375555a2496"} Nov 25 22:05:13 crc kubenswrapper[4931]: I1125 22:05:13.758951 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nqrm9" podStartSLOduration=2.28511329 podStartE2EDuration="4.758930326s" podCreationTimestamp="2025-11-25 22:05:09 +0000 UTC" firstStartedPulling="2025-11-25 22:05:10.695139254 +0000 UTC m=+849.051651383" lastFinishedPulling="2025-11-25 22:05:13.16895626 +0000 UTC m=+851.525468419" observedRunningTime="2025-11-25 22:05:13.755244026 +0000 UTC m=+852.111756155" watchObservedRunningTime="2025-11-25 22:05:13.758930326 +0000 UTC m=+852.115442475" Nov 25 22:05:13 crc kubenswrapper[4931]: I1125 22:05:13.784780 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-kv5fj" podStartSLOduration=2.638937758 podStartE2EDuration="3.784751952s" podCreationTimestamp="2025-11-25 22:05:10 +0000 UTC" firstStartedPulling="2025-11-25 22:05:11.46992222 +0000 UTC m=+849.826434359" lastFinishedPulling="2025-11-25 22:05:12.615736424 +0000 UTC m=+850.972248553" observedRunningTime="2025-11-25 22:05:13.774589898 +0000 UTC m=+852.131102047" watchObservedRunningTime="2025-11-25 22:05:13.784751952 +0000 UTC m=+852.141264091" Nov 25 22:05:20 crc kubenswrapper[4931]: I1125 22:05:20.044248 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:20 crc kubenswrapper[4931]: I1125 22:05:20.044630 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:20 crc kubenswrapper[4931]: I1125 22:05:20.112284 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:20 crc kubenswrapper[4931]: I1125 22:05:20.838397 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:21 crc kubenswrapper[4931]: I1125 22:05:21.025078 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-kv5fj" Nov 25 22:05:21 crc kubenswrapper[4931]: I1125 22:05:21.025457 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-kv5fj" Nov 25 22:05:21 crc kubenswrapper[4931]: I1125 22:05:21.066064 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-kv5fj" Nov 25 22:05:21 crc kubenswrapper[4931]: I1125 22:05:21.688252 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nqrm9"] Nov 25 22:05:21 crc kubenswrapper[4931]: I1125 22:05:21.818423 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-kv5fj" Nov 25 22:05:22 crc kubenswrapper[4931]: I1125 22:05:22.789940 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nqrm9" podUID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerName="registry-server" containerID="cri-o://2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3" gracePeriod=2 Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.212313 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.366972 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-utilities\") pod \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.367042 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2qqm\" (UniqueName: \"kubernetes.io/projected/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-kube-api-access-f2qqm\") pod \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.367124 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-catalog-content\") pod \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\" (UID: \"dd248e3a-0fd2-4616-8b2f-c9e5671650ce\") " Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.368228 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-utilities" (OuterVolumeSpecName: "utilities") pod "dd248e3a-0fd2-4616-8b2f-c9e5671650ce" (UID: "dd248e3a-0fd2-4616-8b2f-c9e5671650ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.373065 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-kube-api-access-f2qqm" (OuterVolumeSpecName: "kube-api-access-f2qqm") pod "dd248e3a-0fd2-4616-8b2f-c9e5671650ce" (UID: "dd248e3a-0fd2-4616-8b2f-c9e5671650ce"). InnerVolumeSpecName "kube-api-access-f2qqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.461584 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd248e3a-0fd2-4616-8b2f-c9e5671650ce" (UID: "dd248e3a-0fd2-4616-8b2f-c9e5671650ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.468993 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.469027 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2qqm\" (UniqueName: \"kubernetes.io/projected/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-kube-api-access-f2qqm\") on node \"crc\" DevicePath \"\"" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.469039 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd248e3a-0fd2-4616-8b2f-c9e5671650ce-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.799258 4931 generic.go:334] "Generic (PLEG): container finished" podID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerID="2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3" exitCode=0 Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.799324 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqrm9" event={"ID":"dd248e3a-0fd2-4616-8b2f-c9e5671650ce","Type":"ContainerDied","Data":"2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3"} Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.799350 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqrm9" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.799382 4931 scope.go:117] "RemoveContainer" containerID="2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.799367 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqrm9" event={"ID":"dd248e3a-0fd2-4616-8b2f-c9e5671650ce","Type":"ContainerDied","Data":"f8bed3e9c13b8d5809a16339b58e1fcaa312280d6edcb26232208817cefab16b"} Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.822860 4931 scope.go:117] "RemoveContainer" containerID="1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.832741 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nqrm9"] Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.838526 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nqrm9"] Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.862127 4931 scope.go:117] "RemoveContainer" containerID="05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.875290 4931 scope.go:117] "RemoveContainer" containerID="2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3" Nov 25 22:05:23 crc kubenswrapper[4931]: E1125 22:05:23.875739 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3\": container with ID starting with 2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3 not found: ID does not exist" containerID="2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.875782 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3"} err="failed to get container status \"2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3\": rpc error: code = NotFound desc = could not find container \"2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3\": container with ID starting with 2cc550ec51bbacc8a35e0b2128bcf948e4f433ee2fd2800b230c9ca12c553be3 not found: ID does not exist" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.875815 4931 scope.go:117] "RemoveContainer" containerID="1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0" Nov 25 22:05:23 crc kubenswrapper[4931]: E1125 22:05:23.876201 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0\": container with ID starting with 1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0 not found: ID does not exist" containerID="1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.876275 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0"} err="failed to get container status \"1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0\": rpc error: code = NotFound desc = could not find container \"1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0\": container with ID starting with 1b9aac3808647cbe77a87d0b6f765a9c4938eae4eb45a49bdbd720b8bf89abb0 not found: ID does not exist" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.876343 4931 scope.go:117] "RemoveContainer" containerID="05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0" Nov 25 22:05:23 crc kubenswrapper[4931]: E1125 22:05:23.876739 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0\": container with ID starting with 05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0 not found: ID does not exist" containerID="05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0" Nov 25 22:05:23 crc kubenswrapper[4931]: I1125 22:05:23.876790 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0"} err="failed to get container status \"05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0\": rpc error: code = NotFound desc = could not find container \"05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0\": container with ID starting with 05f90c90c78f13ac88e3b77414c6965581d76953680b0e853ded10f4f23639b0 not found: ID does not exist" Nov 25 22:05:24 crc kubenswrapper[4931]: I1125 22:05:24.430733 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" path="/var/lib/kubelet/pods/dd248e3a-0fd2-4616-8b2f-c9e5671650ce/volumes" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.754316 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg"] Nov 25 22:05:29 crc kubenswrapper[4931]: E1125 22:05:29.757155 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerName="extract-utilities" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.757599 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerName="extract-utilities" Nov 25 22:05:29 crc kubenswrapper[4931]: E1125 22:05:29.757823 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerName="extract-content" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.758057 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerName="extract-content" Nov 25 22:05:29 crc kubenswrapper[4931]: E1125 22:05:29.758232 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerName="registry-server" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.758384 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerName="registry-server" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.758812 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd248e3a-0fd2-4616-8b2f-c9e5671650ce" containerName="registry-server" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.761134 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.763160 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg"] Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.763225 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrsb7" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.859915 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.859976 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rppld\" (UniqueName: \"kubernetes.io/projected/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-kube-api-access-rppld\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.860036 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.960954 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rppld\" (UniqueName: \"kubernetes.io/projected/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-kube-api-access-rppld\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.961097 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.961208 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.962170 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.962241 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:29 crc kubenswrapper[4931]: I1125 22:05:29.983227 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rppld\" (UniqueName: \"kubernetes.io/projected/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-kube-api-access-rppld\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:30 crc kubenswrapper[4931]: I1125 22:05:30.095391 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:30 crc kubenswrapper[4931]: I1125 22:05:30.503140 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg"] Nov 25 22:05:30 crc kubenswrapper[4931]: I1125 22:05:30.857304 4931 generic.go:334] "Generic (PLEG): container finished" podID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerID="d830c3cca3bc068c21855334c0d92946273c2fd7057fb50f8c908e871ed75c06" exitCode=0 Nov 25 22:05:30 crc kubenswrapper[4931]: I1125 22:05:30.857393 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" event={"ID":"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c","Type":"ContainerDied","Data":"d830c3cca3bc068c21855334c0d92946273c2fd7057fb50f8c908e871ed75c06"} Nov 25 22:05:30 crc kubenswrapper[4931]: I1125 22:05:30.857609 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" event={"ID":"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c","Type":"ContainerStarted","Data":"4afe551b8bac104444eb3dbed1bf900b236bbf816192deb0f60ed7d613955264"} Nov 25 22:05:31 crc kubenswrapper[4931]: I1125 22:05:31.866147 4931 generic.go:334] "Generic (PLEG): container finished" podID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerID="5d5a4c9cbef36225552c1d18faddc2b069612a9dc1f75147868b007141c44b2a" exitCode=0 Nov 25 22:05:31 crc kubenswrapper[4931]: I1125 22:05:31.866213 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" event={"ID":"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c","Type":"ContainerDied","Data":"5d5a4c9cbef36225552c1d18faddc2b069612a9dc1f75147868b007141c44b2a"} Nov 25 22:05:32 crc kubenswrapper[4931]: I1125 22:05:32.878449 4931 generic.go:334] "Generic (PLEG): container finished" podID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerID="1e7f659dad304f752d4bf10e41d06cd6b1ce026da2dc226d03403e579c39bff0" exitCode=0 Nov 25 22:05:32 crc kubenswrapper[4931]: I1125 22:05:32.878532 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" event={"ID":"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c","Type":"ContainerDied","Data":"1e7f659dad304f752d4bf10e41d06cd6b1ce026da2dc226d03403e579c39bff0"} Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.136293 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.320636 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rppld\" (UniqueName: \"kubernetes.io/projected/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-kube-api-access-rppld\") pod \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.320794 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-util\") pod \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.320929 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-bundle\") pod \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\" (UID: \"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c\") " Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.322978 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-bundle" (OuterVolumeSpecName: "bundle") pod "a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" (UID: "a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.325722 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-kube-api-access-rppld" (OuterVolumeSpecName: "kube-api-access-rppld") pod "a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" (UID: "a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c"). InnerVolumeSpecName "kube-api-access-rppld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.354168 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-util" (OuterVolumeSpecName: "util") pod "a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" (UID: "a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.422799 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-util\") on node \"crc\" DevicePath \"\"" Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.422845 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.422858 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rppld\" (UniqueName: \"kubernetes.io/projected/a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c-kube-api-access-rppld\") on node \"crc\" DevicePath \"\"" Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.894901 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" event={"ID":"a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c","Type":"ContainerDied","Data":"4afe551b8bac104444eb3dbed1bf900b236bbf816192deb0f60ed7d613955264"} Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.895279 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4afe551b8bac104444eb3dbed1bf900b236bbf816192deb0f60ed7d613955264" Nov 25 22:05:34 crc kubenswrapper[4931]: I1125 22:05:34.894954 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.179120 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq"] Nov 25 22:05:42 crc kubenswrapper[4931]: E1125 22:05:42.179711 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerName="util" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.179722 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerName="util" Nov 25 22:05:42 crc kubenswrapper[4931]: E1125 22:05:42.179730 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerName="pull" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.179735 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerName="pull" Nov 25 22:05:42 crc kubenswrapper[4931]: E1125 22:05:42.179754 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerName="extract" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.179761 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerName="extract" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.179847 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c" containerName="extract" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.180456 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.182752 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pdm5c" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.183059 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.200128 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq"] Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.328083 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4575978-a78e-42b2-a2dc-1052e5777b04-apiservice-cert\") pod \"infra-operator-controller-manager-6545c6c9d6-7rkzq\" (UID: \"f4575978-a78e-42b2-a2dc-1052e5777b04\") " pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.328608 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4575978-a78e-42b2-a2dc-1052e5777b04-webhook-cert\") pod \"infra-operator-controller-manager-6545c6c9d6-7rkzq\" (UID: \"f4575978-a78e-42b2-a2dc-1052e5777b04\") " pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.328785 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5crj7\" (UniqueName: \"kubernetes.io/projected/f4575978-a78e-42b2-a2dc-1052e5777b04-kube-api-access-5crj7\") pod \"infra-operator-controller-manager-6545c6c9d6-7rkzq\" (UID: \"f4575978-a78e-42b2-a2dc-1052e5777b04\") " pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.430024 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4575978-a78e-42b2-a2dc-1052e5777b04-apiservice-cert\") pod \"infra-operator-controller-manager-6545c6c9d6-7rkzq\" (UID: \"f4575978-a78e-42b2-a2dc-1052e5777b04\") " pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.430132 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4575978-a78e-42b2-a2dc-1052e5777b04-webhook-cert\") pod \"infra-operator-controller-manager-6545c6c9d6-7rkzq\" (UID: \"f4575978-a78e-42b2-a2dc-1052e5777b04\") " pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.430161 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5crj7\" (UniqueName: \"kubernetes.io/projected/f4575978-a78e-42b2-a2dc-1052e5777b04-kube-api-access-5crj7\") pod \"infra-operator-controller-manager-6545c6c9d6-7rkzq\" (UID: \"f4575978-a78e-42b2-a2dc-1052e5777b04\") " pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.436842 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4575978-a78e-42b2-a2dc-1052e5777b04-apiservice-cert\") pod \"infra-operator-controller-manager-6545c6c9d6-7rkzq\" (UID: \"f4575978-a78e-42b2-a2dc-1052e5777b04\") " pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.441579 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4575978-a78e-42b2-a2dc-1052e5777b04-webhook-cert\") pod \"infra-operator-controller-manager-6545c6c9d6-7rkzq\" (UID: \"f4575978-a78e-42b2-a2dc-1052e5777b04\") " pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.456059 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5crj7\" (UniqueName: \"kubernetes.io/projected/f4575978-a78e-42b2-a2dc-1052e5777b04-kube-api-access-5crj7\") pod \"infra-operator-controller-manager-6545c6c9d6-7rkzq\" (UID: \"f4575978-a78e-42b2-a2dc-1052e5777b04\") " pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.496660 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.913220 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq"] Nov 25 22:05:42 crc kubenswrapper[4931]: I1125 22:05:42.964844 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" event={"ID":"f4575978-a78e-42b2-a2dc-1052e5777b04","Type":"ContainerStarted","Data":"9fc5926f8af36403cc65008565832fc2ae23b7b0747d2c95882b1d952bf29077"} Nov 25 22:05:44 crc kubenswrapper[4931]: I1125 22:05:44.986151 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" event={"ID":"f4575978-a78e-42b2-a2dc-1052e5777b04","Type":"ContainerStarted","Data":"49f7a9f920f39948ab4b6feb74172179b41a03f604bebb7cb1d51e2f0adb0762"} Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.103451 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pxkhc"] Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.105377 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.116528 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxkhc"] Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.187060 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfgzd\" (UniqueName: \"kubernetes.io/projected/1f8115aa-27e9-439a-9f8e-dfec2058cec7-kube-api-access-pfgzd\") pod \"community-operators-pxkhc\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.187159 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-utilities\") pod \"community-operators-pxkhc\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.187199 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-catalog-content\") pod \"community-operators-pxkhc\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.288159 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-utilities\") pod \"community-operators-pxkhc\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.288219 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-catalog-content\") pod \"community-operators-pxkhc\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.288273 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfgzd\" (UniqueName: \"kubernetes.io/projected/1f8115aa-27e9-439a-9f8e-dfec2058cec7-kube-api-access-pfgzd\") pod \"community-operators-pxkhc\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.288851 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-utilities\") pod \"community-operators-pxkhc\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.288907 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-catalog-content\") pod \"community-operators-pxkhc\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.308866 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfgzd\" (UniqueName: \"kubernetes.io/projected/1f8115aa-27e9-439a-9f8e-dfec2058cec7-kube-api-access-pfgzd\") pod \"community-operators-pxkhc\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.424718 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:47 crc kubenswrapper[4931]: I1125 22:05:47.775159 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxkhc"] Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.005268 4931 generic.go:334] "Generic (PLEG): container finished" podID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerID="99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3" exitCode=0 Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.005357 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxkhc" event={"ID":"1f8115aa-27e9-439a-9f8e-dfec2058cec7","Type":"ContainerDied","Data":"99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3"} Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.005449 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxkhc" event={"ID":"1f8115aa-27e9-439a-9f8e-dfec2058cec7","Type":"ContainerStarted","Data":"2adda3c39e6a1b30c05e2341410b8afc69359da1484fa83dc629c02adad82d67"} Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.009282 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" event={"ID":"f4575978-a78e-42b2-a2dc-1052e5777b04","Type":"ContainerStarted","Data":"f2b30ffa2b1bcbef8d66dd7138f4b38f51d78f4c1ed551744b873e722a574b1a"} Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.009855 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.045834 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" podStartSLOduration=1.753328404 podStartE2EDuration="6.045817202s" podCreationTimestamp="2025-11-25 22:05:42 +0000 UTC" firstStartedPulling="2025-11-25 22:05:42.930557385 +0000 UTC m=+881.287069514" lastFinishedPulling="2025-11-25 22:05:47.223046193 +0000 UTC m=+885.579558312" observedRunningTime="2025-11-25 22:05:48.045616247 +0000 UTC m=+886.402128376" watchObservedRunningTime="2025-11-25 22:05:48.045817202 +0000 UTC m=+886.402329321" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.210012 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.226829 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.230275 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.230705 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-vn6l5" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.230812 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.232102 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.230859 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.237044 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.241205 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.242337 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.249199 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.253050 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.254222 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.284290 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.404976 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe87b340-3470-4f3e-a69a-e041cdcd24c2-kolla-config\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405025 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-config-data-default\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405062 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405079 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-kolla-config\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405189 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpfwk\" (UniqueName: \"kubernetes.io/projected/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-kube-api-access-wpfwk\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405261 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405287 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe87b340-3470-4f3e-a69a-e041cdcd24c2-operator-scripts\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405337 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cbp2\" (UniqueName: \"kubernetes.io/projected/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-kube-api-access-8cbp2\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405363 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-operator-scripts\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405422 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-config-data-generated\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405448 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-kolla-config\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405466 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe87b340-3470-4f3e-a69a-e041cdcd24c2-config-data-default\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405491 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-config-data-default\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405508 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqpjr\" (UniqueName: \"kubernetes.io/projected/fe87b340-3470-4f3e-a69a-e041cdcd24c2-kube-api-access-cqpjr\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405617 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405700 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe87b340-3470-4f3e-a69a-e041cdcd24c2-config-data-generated\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405743 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.405764 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.507403 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe87b340-3470-4f3e-a69a-e041cdcd24c2-kolla-config\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.507525 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-config-data-default\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.507657 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.507716 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-kolla-config\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.507773 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpfwk\" (UniqueName: \"kubernetes.io/projected/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-kube-api-access-wpfwk\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.507841 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.507966 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe87b340-3470-4f3e-a69a-e041cdcd24c2-operator-scripts\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508062 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-operator-scripts\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508118 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cbp2\" (UniqueName: \"kubernetes.io/projected/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-kube-api-access-8cbp2\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508194 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-config-data-generated\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508235 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508255 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-kolla-config\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508314 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe87b340-3470-4f3e-a69a-e041cdcd24c2-config-data-default\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508392 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-config-data-default\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508452 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpjr\" (UniqueName: \"kubernetes.io/projected/fe87b340-3470-4f3e-a69a-e041cdcd24c2-kube-api-access-cqpjr\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508517 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508598 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe87b340-3470-4f3e-a69a-e041cdcd24c2-config-data-generated\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508658 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508710 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.508791 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-config-data-default\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.509157 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.509459 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.509538 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-config-data-default\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.509697 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-config-data-generated\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.509830 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe87b340-3470-4f3e-a69a-e041cdcd24c2-config-data-generated\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.509888 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.510242 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.510303 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-operator-scripts\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.510443 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-kolla-config\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.510780 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe87b340-3470-4f3e-a69a-e041cdcd24c2-operator-scripts\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.510989 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe87b340-3470-4f3e-a69a-e041cdcd24c2-config-data-default\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.511361 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe87b340-3470-4f3e-a69a-e041cdcd24c2-kolla-config\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.512273 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-kolla-config\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.528128 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.529920 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpfwk\" (UniqueName: \"kubernetes.io/projected/c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a-kube-api-access-wpfwk\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.532184 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-2\" (UID: \"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.536282 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cbp2\" (UniqueName: \"kubernetes.io/projected/56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb-kube-api-access-8cbp2\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.536739 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.540010 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqpjr\" (UniqueName: \"kubernetes.io/projected/fe87b340-3470-4f3e-a69a-e041cdcd24c2-kube-api-access-cqpjr\") pod \"openstack-galera-1\" (UID: \"fe87b340-3470-4f3e-a69a-e041cdcd24c2\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.549530 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.606191 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:05:48 crc kubenswrapper[4931]: I1125 22:05:48.614928 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:05:49 crc kubenswrapper[4931]: I1125 22:05:49.020240 4931 generic.go:334] "Generic (PLEG): container finished" podID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerID="888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2" exitCode=0 Nov 25 22:05:49 crc kubenswrapper[4931]: I1125 22:05:49.020337 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxkhc" event={"ID":"1f8115aa-27e9-439a-9f8e-dfec2058cec7","Type":"ContainerDied","Data":"888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2"} Nov 25 22:05:49 crc kubenswrapper[4931]: I1125 22:05:49.029906 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:05:49 crc kubenswrapper[4931]: I1125 22:05:49.030387 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 22:05:49 crc kubenswrapper[4931]: W1125 22:05:49.068836 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56fd8864_ec73_42e3_8dd9_0d9f5d29d7eb.slice/crio-c2c56a37a15315b5a456c4f347b48ecc543b22ba642c0cc95890ff1d669c048e WatchSource:0}: Error finding container c2c56a37a15315b5a456c4f347b48ecc543b22ba642c0cc95890ff1d669c048e: Status 404 returned error can't find the container with id c2c56a37a15315b5a456c4f347b48ecc543b22ba642c0cc95890ff1d669c048e Nov 25 22:05:49 crc kubenswrapper[4931]: I1125 22:05:49.135128 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 22:05:49 crc kubenswrapper[4931]: W1125 22:05:49.143179 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe87b340_3470_4f3e_a69a_e041cdcd24c2.slice/crio-a2baf1a8f670004a72a706b33b6ec413882b788ed6709d01842f23eafba24835 WatchSource:0}: Error finding container a2baf1a8f670004a72a706b33b6ec413882b788ed6709d01842f23eafba24835: Status 404 returned error can't find the container with id a2baf1a8f670004a72a706b33b6ec413882b788ed6709d01842f23eafba24835 Nov 25 22:05:49 crc kubenswrapper[4931]: I1125 22:05:49.155370 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 22:05:49 crc kubenswrapper[4931]: W1125 22:05:49.162223 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7a1ed06_d8b3_42ef_9fb0_795bbc337b3a.slice/crio-23409cbce19c229205a75296a459beb261511ab18c6e33d0766fa7b093794bd9 WatchSource:0}: Error finding container 23409cbce19c229205a75296a459beb261511ab18c6e33d0766fa7b093794bd9: Status 404 returned error can't find the container with id 23409cbce19c229205a75296a459beb261511ab18c6e33d0766fa7b093794bd9 Nov 25 22:05:50 crc kubenswrapper[4931]: I1125 22:05:50.049239 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb","Type":"ContainerStarted","Data":"c2c56a37a15315b5a456c4f347b48ecc543b22ba642c0cc95890ff1d669c048e"} Nov 25 22:05:50 crc kubenswrapper[4931]: I1125 22:05:50.050129 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a","Type":"ContainerStarted","Data":"23409cbce19c229205a75296a459beb261511ab18c6e33d0766fa7b093794bd9"} Nov 25 22:05:50 crc kubenswrapper[4931]: I1125 22:05:50.062689 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"fe87b340-3470-4f3e-a69a-e041cdcd24c2","Type":"ContainerStarted","Data":"a2baf1a8f670004a72a706b33b6ec413882b788ed6709d01842f23eafba24835"} Nov 25 22:05:50 crc kubenswrapper[4931]: I1125 22:05:50.071280 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxkhc" event={"ID":"1f8115aa-27e9-439a-9f8e-dfec2058cec7","Type":"ContainerStarted","Data":"57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a"} Nov 25 22:05:52 crc kubenswrapper[4931]: I1125 22:05:52.445091 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pxkhc" podStartSLOduration=3.9605805480000003 podStartE2EDuration="5.445074287s" podCreationTimestamp="2025-11-25 22:05:47 +0000 UTC" firstStartedPulling="2025-11-25 22:05:48.008120226 +0000 UTC m=+886.364632365" lastFinishedPulling="2025-11-25 22:05:49.492613975 +0000 UTC m=+887.849126104" observedRunningTime="2025-11-25 22:05:50.093568107 +0000 UTC m=+888.450080226" watchObservedRunningTime="2025-11-25 22:05:52.445074287 +0000 UTC m=+890.801586406" Nov 25 22:05:52 crc kubenswrapper[4931]: I1125 22:05:52.910634 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 22:05:52 crc kubenswrapper[4931]: I1125 22:05:52.911336 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:52 crc kubenswrapper[4931]: I1125 22:05:52.913053 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 25 22:05:52 crc kubenswrapper[4931]: I1125 22:05:52.913363 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-whzws" Nov 25 22:05:52 crc kubenswrapper[4931]: I1125 22:05:52.924243 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.098694 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4ddx\" (UniqueName: \"kubernetes.io/projected/27505da7-a979-42ac-9572-b8a2f6a075de-kube-api-access-t4ddx\") pod \"memcached-0\" (UID: \"27505da7-a979-42ac-9572-b8a2f6a075de\") " pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.098765 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27505da7-a979-42ac-9572-b8a2f6a075de-kolla-config\") pod \"memcached-0\" (UID: \"27505da7-a979-42ac-9572-b8a2f6a075de\") " pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.098792 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27505da7-a979-42ac-9572-b8a2f6a075de-config-data\") pod \"memcached-0\" (UID: \"27505da7-a979-42ac-9572-b8a2f6a075de\") " pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.199780 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27505da7-a979-42ac-9572-b8a2f6a075de-config-data\") pod \"memcached-0\" (UID: \"27505da7-a979-42ac-9572-b8a2f6a075de\") " pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.199896 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4ddx\" (UniqueName: \"kubernetes.io/projected/27505da7-a979-42ac-9572-b8a2f6a075de-kube-api-access-t4ddx\") pod \"memcached-0\" (UID: \"27505da7-a979-42ac-9572-b8a2f6a075de\") " pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.199929 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27505da7-a979-42ac-9572-b8a2f6a075de-kolla-config\") pod \"memcached-0\" (UID: \"27505da7-a979-42ac-9572-b8a2f6a075de\") " pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.200908 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27505da7-a979-42ac-9572-b8a2f6a075de-config-data\") pod \"memcached-0\" (UID: \"27505da7-a979-42ac-9572-b8a2f6a075de\") " pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.200964 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27505da7-a979-42ac-9572-b8a2f6a075de-kolla-config\") pod \"memcached-0\" (UID: \"27505da7-a979-42ac-9572-b8a2f6a075de\") " pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.239706 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4ddx\" (UniqueName: \"kubernetes.io/projected/27505da7-a979-42ac-9572-b8a2f6a075de-kube-api-access-t4ddx\") pod \"memcached-0\" (UID: \"27505da7-a979-42ac-9572-b8a2f6a075de\") " pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.494640 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5f67h"] Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.496347 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.522925 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f67h"] Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.538751 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.605494 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-catalog-content\") pod \"redhat-marketplace-5f67h\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.605574 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-utilities\") pod \"redhat-marketplace-5f67h\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.605631 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqbj6\" (UniqueName: \"kubernetes.io/projected/c35554e5-9c45-4354-bfc4-daae7b9db155-kube-api-access-rqbj6\") pod \"redhat-marketplace-5f67h\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.706829 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-utilities\") pod \"redhat-marketplace-5f67h\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.706908 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqbj6\" (UniqueName: \"kubernetes.io/projected/c35554e5-9c45-4354-bfc4-daae7b9db155-kube-api-access-rqbj6\") pod \"redhat-marketplace-5f67h\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.706946 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-catalog-content\") pod \"redhat-marketplace-5f67h\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.707467 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-catalog-content\") pod \"redhat-marketplace-5f67h\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.707520 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-utilities\") pod \"redhat-marketplace-5f67h\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.735681 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqbj6\" (UniqueName: \"kubernetes.io/projected/c35554e5-9c45-4354-bfc4-daae7b9db155-kube-api-access-rqbj6\") pod \"redhat-marketplace-5f67h\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:53 crc kubenswrapper[4931]: I1125 22:05:53.811580 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.425861 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.426290 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.483197 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.490360 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-t6lmz"] Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.491056 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.498078 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-j5k42" Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.506419 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-t6lmz"] Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.670714 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2mj5\" (UniqueName: \"kubernetes.io/projected/ad908caf-8931-44bb-a5b3-a59952dfc01a-kube-api-access-w2mj5\") pod \"rabbitmq-cluster-operator-index-t6lmz\" (UID: \"ad908caf-8931-44bb-a5b3-a59952dfc01a\") " pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.772409 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2mj5\" (UniqueName: \"kubernetes.io/projected/ad908caf-8931-44bb-a5b3-a59952dfc01a-kube-api-access-w2mj5\") pod \"rabbitmq-cluster-operator-index-t6lmz\" (UID: \"ad908caf-8931-44bb-a5b3-a59952dfc01a\") " pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.792191 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2mj5\" (UniqueName: \"kubernetes.io/projected/ad908caf-8931-44bb-a5b3-a59952dfc01a-kube-api-access-w2mj5\") pod \"rabbitmq-cluster-operator-index-t6lmz\" (UID: \"ad908caf-8931-44bb-a5b3-a59952dfc01a\") " pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" Nov 25 22:05:57 crc kubenswrapper[4931]: I1125 22:05:57.882335 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" Nov 25 22:05:58 crc kubenswrapper[4931]: I1125 22:05:58.165550 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:05:58 crc kubenswrapper[4931]: I1125 22:05:58.712718 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-t6lmz"] Nov 25 22:05:58 crc kubenswrapper[4931]: I1125 22:05:58.806977 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f67h"] Nov 25 22:05:58 crc kubenswrapper[4931]: W1125 22:05:58.814066 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc35554e5_9c45_4354_bfc4_daae7b9db155.slice/crio-d05b6460244867987189296d013e0413c03e35f1182bf64a5fccfcb41c3d7922 WatchSource:0}: Error finding container d05b6460244867987189296d013e0413c03e35f1182bf64a5fccfcb41c3d7922: Status 404 returned error can't find the container with id d05b6460244867987189296d013e0413c03e35f1182bf64a5fccfcb41c3d7922 Nov 25 22:05:58 crc kubenswrapper[4931]: I1125 22:05:58.931470 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 22:05:58 crc kubenswrapper[4931]: W1125 22:05:58.939063 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27505da7_a979_42ac_9572_b8a2f6a075de.slice/crio-bbeb91b5b55fb0b61d4c48096b8e0689c50e4902528729a93cb3c6329ae4f663 WatchSource:0}: Error finding container bbeb91b5b55fb0b61d4c48096b8e0689c50e4902528729a93cb3c6329ae4f663: Status 404 returned error can't find the container with id bbeb91b5b55fb0b61d4c48096b8e0689c50e4902528729a93cb3c6329ae4f663 Nov 25 22:05:59 crc kubenswrapper[4931]: I1125 22:05:59.134129 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a","Type":"ContainerStarted","Data":"813286e05e6598eb2f49159f737ae938ef8abb0b6cd95123243def7040c6d3a1"} Nov 25 22:05:59 crc kubenswrapper[4931]: I1125 22:05:59.136398 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"fe87b340-3470-4f3e-a69a-e041cdcd24c2","Type":"ContainerStarted","Data":"80d4e9c591a785c032ba085e0fc8985190925e50d233a4d798773858c386c355"} Nov 25 22:05:59 crc kubenswrapper[4931]: I1125 22:05:59.138355 4931 generic.go:334] "Generic (PLEG): container finished" podID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerID="d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa" exitCode=0 Nov 25 22:05:59 crc kubenswrapper[4931]: I1125 22:05:59.138435 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f67h" event={"ID":"c35554e5-9c45-4354-bfc4-daae7b9db155","Type":"ContainerDied","Data":"d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa"} Nov 25 22:05:59 crc kubenswrapper[4931]: I1125 22:05:59.138520 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f67h" event={"ID":"c35554e5-9c45-4354-bfc4-daae7b9db155","Type":"ContainerStarted","Data":"d05b6460244867987189296d013e0413c03e35f1182bf64a5fccfcb41c3d7922"} Nov 25 22:05:59 crc kubenswrapper[4931]: I1125 22:05:59.140859 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb","Type":"ContainerStarted","Data":"5b42531191aa0fd7625c27ba1991272f2285ba8a95ccba2c61989df1689aec63"} Nov 25 22:05:59 crc kubenswrapper[4931]: I1125 22:05:59.144054 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" event={"ID":"ad908caf-8931-44bb-a5b3-a59952dfc01a","Type":"ContainerStarted","Data":"247d173afbfd4f201ec29912f70e3e802424f813dc3eaa68ac7911dcb8b7ecd3"} Nov 25 22:05:59 crc kubenswrapper[4931]: I1125 22:05:59.145846 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"27505da7-a979-42ac-9572-b8a2f6a075de","Type":"ContainerStarted","Data":"bbeb91b5b55fb0b61d4c48096b8e0689c50e4902528729a93cb3c6329ae4f663"} Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.087648 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxkhc"] Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.088453 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pxkhc" podUID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerName="registry-server" containerID="cri-o://57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a" gracePeriod=2 Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.811586 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.883568 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-utilities\") pod \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.883640 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfgzd\" (UniqueName: \"kubernetes.io/projected/1f8115aa-27e9-439a-9f8e-dfec2058cec7-kube-api-access-pfgzd\") pod \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.884456 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-utilities" (OuterVolumeSpecName: "utilities") pod "1f8115aa-27e9-439a-9f8e-dfec2058cec7" (UID: "1f8115aa-27e9-439a-9f8e-dfec2058cec7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.889407 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f8115aa-27e9-439a-9f8e-dfec2058cec7-kube-api-access-pfgzd" (OuterVolumeSpecName: "kube-api-access-pfgzd") pod "1f8115aa-27e9-439a-9f8e-dfec2058cec7" (UID: "1f8115aa-27e9-439a-9f8e-dfec2058cec7"). InnerVolumeSpecName "kube-api-access-pfgzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.984317 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-catalog-content\") pod \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\" (UID: \"1f8115aa-27e9-439a-9f8e-dfec2058cec7\") " Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.984666 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:03 crc kubenswrapper[4931]: I1125 22:06:03.984680 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfgzd\" (UniqueName: \"kubernetes.io/projected/1f8115aa-27e9-439a-9f8e-dfec2058cec7-kube-api-access-pfgzd\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.031498 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f8115aa-27e9-439a-9f8e-dfec2058cec7" (UID: "1f8115aa-27e9-439a-9f8e-dfec2058cec7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.085842 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f8115aa-27e9-439a-9f8e-dfec2058cec7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.217617 4931 generic.go:334] "Generic (PLEG): container finished" podID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerID="57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a" exitCode=0 Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.217719 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxkhc" Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.217750 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxkhc" event={"ID":"1f8115aa-27e9-439a-9f8e-dfec2058cec7","Type":"ContainerDied","Data":"57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a"} Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.218138 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxkhc" event={"ID":"1f8115aa-27e9-439a-9f8e-dfec2058cec7","Type":"ContainerDied","Data":"2adda3c39e6a1b30c05e2341410b8afc69359da1484fa83dc629c02adad82d67"} Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.218169 4931 scope.go:117] "RemoveContainer" containerID="57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a" Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.254031 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxkhc"] Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.260374 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pxkhc"] Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.431302 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" path="/var/lib/kubelet/pods/1f8115aa-27e9-439a-9f8e-dfec2058cec7/volumes" Nov 25 22:06:04 crc kubenswrapper[4931]: I1125 22:06:04.576734 4931 scope.go:117] "RemoveContainer" containerID="888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2" Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.225775 4931 generic.go:334] "Generic (PLEG): container finished" podID="56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb" containerID="5b42531191aa0fd7625c27ba1991272f2285ba8a95ccba2c61989df1689aec63" exitCode=0 Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.225835 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb","Type":"ContainerDied","Data":"5b42531191aa0fd7625c27ba1991272f2285ba8a95ccba2c61989df1689aec63"} Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.231440 4931 generic.go:334] "Generic (PLEG): container finished" podID="c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a" containerID="813286e05e6598eb2f49159f737ae938ef8abb0b6cd95123243def7040c6d3a1" exitCode=0 Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.231526 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a","Type":"ContainerDied","Data":"813286e05e6598eb2f49159f737ae938ef8abb0b6cd95123243def7040c6d3a1"} Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.237088 4931 generic.go:334] "Generic (PLEG): container finished" podID="fe87b340-3470-4f3e-a69a-e041cdcd24c2" containerID="80d4e9c591a785c032ba085e0fc8985190925e50d233a4d798773858c386c355" exitCode=0 Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.237140 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"fe87b340-3470-4f3e-a69a-e041cdcd24c2","Type":"ContainerDied","Data":"80d4e9c591a785c032ba085e0fc8985190925e50d233a4d798773858c386c355"} Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.306722 4931 scope.go:117] "RemoveContainer" containerID="99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3" Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.386997 4931 scope.go:117] "RemoveContainer" containerID="57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a" Nov 25 22:06:05 crc kubenswrapper[4931]: E1125 22:06:05.387674 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a\": container with ID starting with 57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a not found: ID does not exist" containerID="57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a" Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.387720 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a"} err="failed to get container status \"57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a\": rpc error: code = NotFound desc = could not find container \"57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a\": container with ID starting with 57abd6e80227d4e5fe8e2ecb2326fdd390ad74d12ccfedd7f9395884723c520a not found: ID does not exist" Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.387745 4931 scope.go:117] "RemoveContainer" containerID="888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2" Nov 25 22:06:05 crc kubenswrapper[4931]: E1125 22:06:05.388097 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2\": container with ID starting with 888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2 not found: ID does not exist" containerID="888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2" Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.388147 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2"} err="failed to get container status \"888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2\": rpc error: code = NotFound desc = could not find container \"888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2\": container with ID starting with 888224e1baa53b5fe69e4e17f86e27a0be1341846394d8f7311a82433fad3fb2 not found: ID does not exist" Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.388177 4931 scope.go:117] "RemoveContainer" containerID="99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3" Nov 25 22:06:05 crc kubenswrapper[4931]: E1125 22:06:05.388548 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3\": container with ID starting with 99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3 not found: ID does not exist" containerID="99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3" Nov 25 22:06:05 crc kubenswrapper[4931]: I1125 22:06:05.388634 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3"} err="failed to get container status \"99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3\": rpc error: code = NotFound desc = could not find container \"99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3\": container with ID starting with 99f2090bf844407b78bfa1f8f41a811b530958989402e1c28857645d3cbd8be3 not found: ID does not exist" Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.251130 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"fe87b340-3470-4f3e-a69a-e041cdcd24c2","Type":"ContainerStarted","Data":"a0c306c3928d9d33a9a0ec6a90db84e5523b95048c177397c417cdc3935ba34c"} Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.256060 4931 generic.go:334] "Generic (PLEG): container finished" podID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerID="456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392" exitCode=0 Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.256186 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f67h" event={"ID":"c35554e5-9c45-4354-bfc4-daae7b9db155","Type":"ContainerDied","Data":"456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392"} Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.261400 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb","Type":"ContainerStarted","Data":"0e551fd0d1f8ffd7362ed9d85e077262b14804d4719f1cec986cddb4b1d60a4d"} Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.264852 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" event={"ID":"ad908caf-8931-44bb-a5b3-a59952dfc01a","Type":"ContainerStarted","Data":"a20fa32468b206e35fb48768f694b4817123e8b074aefcd8aa8ce36403a4437a"} Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.268572 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"27505da7-a979-42ac-9572-b8a2f6a075de","Type":"ContainerStarted","Data":"044617f78a8319f94a737525b48e512c0138cdf5983b4100126837f6fa4c85db"} Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.268708 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.273489 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a","Type":"ContainerStarted","Data":"5fa2a2e432e421ee90e7b5ef54ec8164ef96050e05b477bdcd5bc7f5c7f587a3"} Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.291817 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=9.98529491 podStartE2EDuration="19.291790517s" podCreationTimestamp="2025-11-25 22:05:47 +0000 UTC" firstStartedPulling="2025-11-25 22:05:49.145627276 +0000 UTC m=+887.502139395" lastFinishedPulling="2025-11-25 22:05:58.452122883 +0000 UTC m=+896.808635002" observedRunningTime="2025-11-25 22:06:06.284534112 +0000 UTC m=+904.641046261" watchObservedRunningTime="2025-11-25 22:06:06.291790517 +0000 UTC m=+904.648302676" Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.344518 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=10.06366515 podStartE2EDuration="19.344474926s" podCreationTimestamp="2025-11-25 22:05:47 +0000 UTC" firstStartedPulling="2025-11-25 22:05:49.164441782 +0000 UTC m=+887.520953901" lastFinishedPulling="2025-11-25 22:05:58.445251558 +0000 UTC m=+896.801763677" observedRunningTime="2025-11-25 22:06:06.338197987 +0000 UTC m=+904.694710116" watchObservedRunningTime="2025-11-25 22:06:06.344474926 +0000 UTC m=+904.700987055" Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.370700 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=9.958851418 podStartE2EDuration="19.370682903s" podCreationTimestamp="2025-11-25 22:05:47 +0000 UTC" firstStartedPulling="2025-11-25 22:05:49.071634622 +0000 UTC m=+887.428146741" lastFinishedPulling="2025-11-25 22:05:58.483466087 +0000 UTC m=+896.839978226" observedRunningTime="2025-11-25 22:06:06.367418395 +0000 UTC m=+904.723930594" watchObservedRunningTime="2025-11-25 22:06:06.370682903 +0000 UTC m=+904.727195022" Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.400084 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=9.193803045 podStartE2EDuration="14.400047994s" podCreationTimestamp="2025-11-25 22:05:52 +0000 UTC" firstStartedPulling="2025-11-25 22:05:58.941340014 +0000 UTC m=+897.297852133" lastFinishedPulling="2025-11-25 22:06:04.147584963 +0000 UTC m=+902.504097082" observedRunningTime="2025-11-25 22:06:06.388939454 +0000 UTC m=+904.745451633" watchObservedRunningTime="2025-11-25 22:06:06.400047994 +0000 UTC m=+904.756560193" Nov 25 22:06:06 crc kubenswrapper[4931]: I1125 22:06:06.417287 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" podStartSLOduration=2.7486237559999998 podStartE2EDuration="9.417270698s" podCreationTimestamp="2025-11-25 22:05:57 +0000 UTC" firstStartedPulling="2025-11-25 22:05:58.71997423 +0000 UTC m=+897.076486349" lastFinishedPulling="2025-11-25 22:06:05.388621182 +0000 UTC m=+903.745133291" observedRunningTime="2025-11-25 22:06:06.413536238 +0000 UTC m=+904.770048367" watchObservedRunningTime="2025-11-25 22:06:06.417270698 +0000 UTC m=+904.773782827" Nov 25 22:06:07 crc kubenswrapper[4931]: I1125 22:06:07.283629 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f67h" event={"ID":"c35554e5-9c45-4354-bfc4-daae7b9db155","Type":"ContainerStarted","Data":"a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239"} Nov 25 22:06:07 crc kubenswrapper[4931]: I1125 22:06:07.313227 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5f67h" podStartSLOduration=6.698373951 podStartE2EDuration="14.313202048s" podCreationTimestamp="2025-11-25 22:05:53 +0000 UTC" firstStartedPulling="2025-11-25 22:05:59.140322695 +0000 UTC m=+897.496834824" lastFinishedPulling="2025-11-25 22:06:06.755150792 +0000 UTC m=+905.111662921" observedRunningTime="2025-11-25 22:06:07.307270818 +0000 UTC m=+905.663782967" watchObservedRunningTime="2025-11-25 22:06:07.313202048 +0000 UTC m=+905.669714197" Nov 25 22:06:07 crc kubenswrapper[4931]: I1125 22:06:07.882312 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" Nov 25 22:06:07 crc kubenswrapper[4931]: I1125 22:06:07.882595 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" Nov 25 22:06:07 crc kubenswrapper[4931]: I1125 22:06:07.923434 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" Nov 25 22:06:08 crc kubenswrapper[4931]: I1125 22:06:08.549606 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:06:08 crc kubenswrapper[4931]: I1125 22:06:08.549660 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:06:08 crc kubenswrapper[4931]: I1125 22:06:08.606617 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:06:08 crc kubenswrapper[4931]: I1125 22:06:08.606699 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:06:08 crc kubenswrapper[4931]: I1125 22:06:08.615452 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:06:08 crc kubenswrapper[4931]: I1125 22:06:08.615504 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:06:13 crc kubenswrapper[4931]: I1125 22:06:13.540210 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 25 22:06:13 crc kubenswrapper[4931]: I1125 22:06:13.812275 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:06:13 crc kubenswrapper[4931]: I1125 22:06:13.812588 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:06:13 crc kubenswrapper[4931]: I1125 22:06:13.895016 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:06:14 crc kubenswrapper[4931]: I1125 22:06:14.391351 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:06:14 crc kubenswrapper[4931]: I1125 22:06:14.842860 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:06:14 crc kubenswrapper[4931]: I1125 22:06:14.943360 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 22:06:17 crc kubenswrapper[4931]: I1125 22:06:17.688143 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f67h"] Nov 25 22:06:17 crc kubenswrapper[4931]: I1125 22:06:17.688581 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5f67h" podUID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerName="registry-server" containerID="cri-o://a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239" gracePeriod=2 Nov 25 22:06:17 crc kubenswrapper[4931]: I1125 22:06:17.942184 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-t6lmz" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.135526 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.210974 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqbj6\" (UniqueName: \"kubernetes.io/projected/c35554e5-9c45-4354-bfc4-daae7b9db155-kube-api-access-rqbj6\") pod \"c35554e5-9c45-4354-bfc4-daae7b9db155\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.211095 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-catalog-content\") pod \"c35554e5-9c45-4354-bfc4-daae7b9db155\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.211161 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-utilities\") pod \"c35554e5-9c45-4354-bfc4-daae7b9db155\" (UID: \"c35554e5-9c45-4354-bfc4-daae7b9db155\") " Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.212533 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-utilities" (OuterVolumeSpecName: "utilities") pod "c35554e5-9c45-4354-bfc4-daae7b9db155" (UID: "c35554e5-9c45-4354-bfc4-daae7b9db155"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.222098 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c35554e5-9c45-4354-bfc4-daae7b9db155-kube-api-access-rqbj6" (OuterVolumeSpecName: "kube-api-access-rqbj6") pod "c35554e5-9c45-4354-bfc4-daae7b9db155" (UID: "c35554e5-9c45-4354-bfc4-daae7b9db155"). InnerVolumeSpecName "kube-api-access-rqbj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.232801 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c35554e5-9c45-4354-bfc4-daae7b9db155" (UID: "c35554e5-9c45-4354-bfc4-daae7b9db155"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.312339 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.312379 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqbj6\" (UniqueName: \"kubernetes.io/projected/c35554e5-9c45-4354-bfc4-daae7b9db155-kube-api-access-rqbj6\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.312393 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c35554e5-9c45-4354-bfc4-daae7b9db155-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.371920 4931 generic.go:334] "Generic (PLEG): container finished" podID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerID="a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239" exitCode=0 Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.371970 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f67h" event={"ID":"c35554e5-9c45-4354-bfc4-daae7b9db155","Type":"ContainerDied","Data":"a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239"} Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.371982 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f67h" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.371999 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f67h" event={"ID":"c35554e5-9c45-4354-bfc4-daae7b9db155","Type":"ContainerDied","Data":"d05b6460244867987189296d013e0413c03e35f1182bf64a5fccfcb41c3d7922"} Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.372019 4931 scope.go:117] "RemoveContainer" containerID="a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.400643 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f67h"] Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.404343 4931 scope.go:117] "RemoveContainer" containerID="456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.405070 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f67h"] Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.433201 4931 scope.go:117] "RemoveContainer" containerID="d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.436225 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c35554e5-9c45-4354-bfc4-daae7b9db155" path="/var/lib/kubelet/pods/c35554e5-9c45-4354-bfc4-daae7b9db155/volumes" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.454315 4931 scope.go:117] "RemoveContainer" containerID="a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239" Nov 25 22:06:18 crc kubenswrapper[4931]: E1125 22:06:18.454923 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239\": container with ID starting with a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239 not found: ID does not exist" containerID="a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.454966 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239"} err="failed to get container status \"a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239\": rpc error: code = NotFound desc = could not find container \"a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239\": container with ID starting with a4551f52d98b4d26546d44feb46237e39e0dbfc445f396ee1eb6fe9c1e059239 not found: ID does not exist" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.454997 4931 scope.go:117] "RemoveContainer" containerID="456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392" Nov 25 22:06:18 crc kubenswrapper[4931]: E1125 22:06:18.455485 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392\": container with ID starting with 456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392 not found: ID does not exist" containerID="456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.455504 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392"} err="failed to get container status \"456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392\": rpc error: code = NotFound desc = could not find container \"456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392\": container with ID starting with 456e481d0a2aa4d45211c5755e890967f497ceb7438788cdfa9329352633a392 not found: ID does not exist" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.455516 4931 scope.go:117] "RemoveContainer" containerID="d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa" Nov 25 22:06:18 crc kubenswrapper[4931]: E1125 22:06:18.455993 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa\": container with ID starting with d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa not found: ID does not exist" containerID="d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa" Nov 25 22:06:18 crc kubenswrapper[4931]: I1125 22:06:18.456012 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa"} err="failed to get container status \"d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa\": rpc error: code = NotFound desc = could not find container \"d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa\": container with ID starting with d7fd5a3ac96d150d6f5ba107a08803514204ae86a07ccb74c651333bcd33c8aa not found: ID does not exist" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.949095 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6"] Nov 25 22:06:27 crc kubenswrapper[4931]: E1125 22:06:27.949844 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerName="extract-content" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.949860 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerName="extract-content" Nov 25 22:06:27 crc kubenswrapper[4931]: E1125 22:06:27.949901 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerName="registry-server" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.949910 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerName="registry-server" Nov 25 22:06:27 crc kubenswrapper[4931]: E1125 22:06:27.949925 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerName="extract-utilities" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.949934 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerName="extract-utilities" Nov 25 22:06:27 crc kubenswrapper[4931]: E1125 22:06:27.949944 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerName="extract-content" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.949953 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerName="extract-content" Nov 25 22:06:27 crc kubenswrapper[4931]: E1125 22:06:27.949968 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerName="registry-server" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.949976 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerName="registry-server" Nov 25 22:06:27 crc kubenswrapper[4931]: E1125 22:06:27.949988 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerName="extract-utilities" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.949996 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerName="extract-utilities" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.950127 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="c35554e5-9c45-4354-bfc4-daae7b9db155" containerName="registry-server" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.950144 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f8115aa-27e9-439a-9f8e-dfec2058cec7" containerName="registry-server" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.951136 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.953563 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrsb7" Nov 25 22:06:27 crc kubenswrapper[4931]: I1125 22:06:27.964656 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6"] Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.050931 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.051237 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6p8n\" (UniqueName: \"kubernetes.io/projected/b979640a-23b5-48df-96be-82bc4630d172-kube-api-access-d6p8n\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.051303 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.153228 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6p8n\" (UniqueName: \"kubernetes.io/projected/b979640a-23b5-48df-96be-82bc4630d172-kube-api-access-d6p8n\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.153320 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.153370 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.153832 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.154110 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.171469 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6p8n\" (UniqueName: \"kubernetes.io/projected/b979640a-23b5-48df-96be-82bc4630d172-kube-api-access-d6p8n\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.266508 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.580558 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.685573 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.709556 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6"] Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.905339 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tglrb"] Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.907164 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.910466 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tglrb"] Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.964727 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-utilities\") pod \"certified-operators-tglrb\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.964836 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg4kf\" (UniqueName: \"kubernetes.io/projected/a3f4937e-64ff-4f9a-addd-0d1f436252d1-kube-api-access-tg4kf\") pod \"certified-operators-tglrb\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:28 crc kubenswrapper[4931]: I1125 22:06:28.964862 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-catalog-content\") pod \"certified-operators-tglrb\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.065916 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg4kf\" (UniqueName: \"kubernetes.io/projected/a3f4937e-64ff-4f9a-addd-0d1f436252d1-kube-api-access-tg4kf\") pod \"certified-operators-tglrb\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.065987 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-catalog-content\") pod \"certified-operators-tglrb\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.066068 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-utilities\") pod \"certified-operators-tglrb\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.066590 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-utilities\") pod \"certified-operators-tglrb\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.066852 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-catalog-content\") pod \"certified-operators-tglrb\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.088000 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg4kf\" (UniqueName: \"kubernetes.io/projected/a3f4937e-64ff-4f9a-addd-0d1f436252d1-kube-api-access-tg4kf\") pod \"certified-operators-tglrb\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.236371 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.432405 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tglrb"] Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.464913 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tglrb" event={"ID":"a3f4937e-64ff-4f9a-addd-0d1f436252d1","Type":"ContainerStarted","Data":"ecec3c5f2f4ad9043c4ff0bcfe360e5e19fd49b0453ec1b201a74faafe3ed17d"} Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.466458 4931 generic.go:334] "Generic (PLEG): container finished" podID="b979640a-23b5-48df-96be-82bc4630d172" containerID="dc07d3cd9190dd3a7ef4001e6459b42c1666850aaff4a300343e101cb17e0b7f" exitCode=0 Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.466557 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" event={"ID":"b979640a-23b5-48df-96be-82bc4630d172","Type":"ContainerDied","Data":"dc07d3cd9190dd3a7ef4001e6459b42c1666850aaff4a300343e101cb17e0b7f"} Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.466615 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" event={"ID":"b979640a-23b5-48df-96be-82bc4630d172","Type":"ContainerStarted","Data":"ec5a443412973dbc3ae3d2c6f3540da97f729e2799c897b37b5ac66384531a87"} Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.600855 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:06:29 crc kubenswrapper[4931]: I1125 22:06:29.676672 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 22:06:30 crc kubenswrapper[4931]: I1125 22:06:30.476362 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" event={"ID":"b979640a-23b5-48df-96be-82bc4630d172","Type":"ContainerStarted","Data":"5667672f140014d1320c65acf7cf5ff214c0976194d36db613d2ec623e87dd0f"} Nov 25 22:06:30 crc kubenswrapper[4931]: I1125 22:06:30.479259 4931 generic.go:334] "Generic (PLEG): container finished" podID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerID="5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6" exitCode=0 Nov 25 22:06:30 crc kubenswrapper[4931]: I1125 22:06:30.479377 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tglrb" event={"ID":"a3f4937e-64ff-4f9a-addd-0d1f436252d1","Type":"ContainerDied","Data":"5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6"} Nov 25 22:06:31 crc kubenswrapper[4931]: I1125 22:06:31.488705 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tglrb" event={"ID":"a3f4937e-64ff-4f9a-addd-0d1f436252d1","Type":"ContainerStarted","Data":"d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3"} Nov 25 22:06:31 crc kubenswrapper[4931]: I1125 22:06:31.491096 4931 generic.go:334] "Generic (PLEG): container finished" podID="b979640a-23b5-48df-96be-82bc4630d172" containerID="5667672f140014d1320c65acf7cf5ff214c0976194d36db613d2ec623e87dd0f" exitCode=0 Nov 25 22:06:31 crc kubenswrapper[4931]: I1125 22:06:31.491146 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" event={"ID":"b979640a-23b5-48df-96be-82bc4630d172","Type":"ContainerDied","Data":"5667672f140014d1320c65acf7cf5ff214c0976194d36db613d2ec623e87dd0f"} Nov 25 22:06:32 crc kubenswrapper[4931]: I1125 22:06:32.501752 4931 generic.go:334] "Generic (PLEG): container finished" podID="b979640a-23b5-48df-96be-82bc4630d172" containerID="0ea668394fe14393f0d0f5cd14b21c496031181b0286ff921d0f396dceebba99" exitCode=0 Nov 25 22:06:32 crc kubenswrapper[4931]: I1125 22:06:32.501851 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" event={"ID":"b979640a-23b5-48df-96be-82bc4630d172","Type":"ContainerDied","Data":"0ea668394fe14393f0d0f5cd14b21c496031181b0286ff921d0f396dceebba99"} Nov 25 22:06:32 crc kubenswrapper[4931]: I1125 22:06:32.504408 4931 generic.go:334] "Generic (PLEG): container finished" podID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerID="d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3" exitCode=0 Nov 25 22:06:32 crc kubenswrapper[4931]: I1125 22:06:32.504451 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tglrb" event={"ID":"a3f4937e-64ff-4f9a-addd-0d1f436252d1","Type":"ContainerDied","Data":"d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3"} Nov 25 22:06:33 crc kubenswrapper[4931]: I1125 22:06:33.512846 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tglrb" event={"ID":"a3f4937e-64ff-4f9a-addd-0d1f436252d1","Type":"ContainerStarted","Data":"1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e"} Nov 25 22:06:33 crc kubenswrapper[4931]: I1125 22:06:33.533183 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tglrb" podStartSLOduration=3.091145671 podStartE2EDuration="5.533167205s" podCreationTimestamp="2025-11-25 22:06:28 +0000 UTC" firstStartedPulling="2025-11-25 22:06:30.481358969 +0000 UTC m=+928.837871098" lastFinishedPulling="2025-11-25 22:06:32.923380483 +0000 UTC m=+931.279892632" observedRunningTime="2025-11-25 22:06:33.529858055 +0000 UTC m=+931.886370194" watchObservedRunningTime="2025-11-25 22:06:33.533167205 +0000 UTC m=+931.889679324" Nov 25 22:06:33 crc kubenswrapper[4931]: I1125 22:06:33.854467 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:33 crc kubenswrapper[4931]: I1125 22:06:33.946233 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-bundle\") pod \"b979640a-23b5-48df-96be-82bc4630d172\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " Nov 25 22:06:33 crc kubenswrapper[4931]: I1125 22:06:33.946484 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6p8n\" (UniqueName: \"kubernetes.io/projected/b979640a-23b5-48df-96be-82bc4630d172-kube-api-access-d6p8n\") pod \"b979640a-23b5-48df-96be-82bc4630d172\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " Nov 25 22:06:33 crc kubenswrapper[4931]: I1125 22:06:33.946669 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-util\") pod \"b979640a-23b5-48df-96be-82bc4630d172\" (UID: \"b979640a-23b5-48df-96be-82bc4630d172\") " Nov 25 22:06:33 crc kubenswrapper[4931]: I1125 22:06:33.947704 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-bundle" (OuterVolumeSpecName: "bundle") pod "b979640a-23b5-48df-96be-82bc4630d172" (UID: "b979640a-23b5-48df-96be-82bc4630d172"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:06:33 crc kubenswrapper[4931]: I1125 22:06:33.954857 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b979640a-23b5-48df-96be-82bc4630d172-kube-api-access-d6p8n" (OuterVolumeSpecName: "kube-api-access-d6p8n") pod "b979640a-23b5-48df-96be-82bc4630d172" (UID: "b979640a-23b5-48df-96be-82bc4630d172"). InnerVolumeSpecName "kube-api-access-d6p8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:06:33 crc kubenswrapper[4931]: I1125 22:06:33.964017 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-util" (OuterVolumeSpecName: "util") pod "b979640a-23b5-48df-96be-82bc4630d172" (UID: "b979640a-23b5-48df-96be-82bc4630d172"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:06:34 crc kubenswrapper[4931]: I1125 22:06:34.048816 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:34 crc kubenswrapper[4931]: I1125 22:06:34.049310 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6p8n\" (UniqueName: \"kubernetes.io/projected/b979640a-23b5-48df-96be-82bc4630d172-kube-api-access-d6p8n\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:34 crc kubenswrapper[4931]: I1125 22:06:34.049496 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b979640a-23b5-48df-96be-82bc4630d172-util\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:34 crc kubenswrapper[4931]: I1125 22:06:34.523011 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" event={"ID":"b979640a-23b5-48df-96be-82bc4630d172","Type":"ContainerDied","Data":"ec5a443412973dbc3ae3d2c6f3540da97f729e2799c897b37b5ac66384531a87"} Nov 25 22:06:34 crc kubenswrapper[4931]: I1125 22:06:34.523071 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec5a443412973dbc3ae3d2c6f3540da97f729e2799c897b37b5ac66384531a87" Nov 25 22:06:34 crc kubenswrapper[4931]: I1125 22:06:34.523035 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6" Nov 25 22:06:39 crc kubenswrapper[4931]: I1125 22:06:39.237473 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:39 crc kubenswrapper[4931]: I1125 22:06:39.237977 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:39 crc kubenswrapper[4931]: I1125 22:06:39.289315 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:39 crc kubenswrapper[4931]: I1125 22:06:39.607676 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:40 crc kubenswrapper[4931]: I1125 22:06:40.883828 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tglrb"] Nov 25 22:06:41 crc kubenswrapper[4931]: I1125 22:06:41.577905 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tglrb" podUID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerName="registry-server" containerID="cri-o://1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e" gracePeriod=2 Nov 25 22:06:41 crc kubenswrapper[4931]: I1125 22:06:41.936067 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.062902 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-utilities\") pod \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.062987 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg4kf\" (UniqueName: \"kubernetes.io/projected/a3f4937e-64ff-4f9a-addd-0d1f436252d1-kube-api-access-tg4kf\") pod \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.063052 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-catalog-content\") pod \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\" (UID: \"a3f4937e-64ff-4f9a-addd-0d1f436252d1\") " Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.064400 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-utilities" (OuterVolumeSpecName: "utilities") pod "a3f4937e-64ff-4f9a-addd-0d1f436252d1" (UID: "a3f4937e-64ff-4f9a-addd-0d1f436252d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.070782 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3f4937e-64ff-4f9a-addd-0d1f436252d1-kube-api-access-tg4kf" (OuterVolumeSpecName: "kube-api-access-tg4kf") pod "a3f4937e-64ff-4f9a-addd-0d1f436252d1" (UID: "a3f4937e-64ff-4f9a-addd-0d1f436252d1"). InnerVolumeSpecName "kube-api-access-tg4kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.164096 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.164135 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg4kf\" (UniqueName: \"kubernetes.io/projected/a3f4937e-64ff-4f9a-addd-0d1f436252d1-kube-api-access-tg4kf\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.585087 4931 generic.go:334] "Generic (PLEG): container finished" podID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerID="1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e" exitCode=0 Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.585124 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tglrb" event={"ID":"a3f4937e-64ff-4f9a-addd-0d1f436252d1","Type":"ContainerDied","Data":"1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e"} Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.585149 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tglrb" event={"ID":"a3f4937e-64ff-4f9a-addd-0d1f436252d1","Type":"ContainerDied","Data":"ecec3c5f2f4ad9043c4ff0bcfe360e5e19fd49b0453ec1b201a74faafe3ed17d"} Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.585164 4931 scope.go:117] "RemoveContainer" containerID="1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.585256 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tglrb" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.608441 4931 scope.go:117] "RemoveContainer" containerID="d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.638253 4931 scope.go:117] "RemoveContainer" containerID="5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.653572 4931 scope.go:117] "RemoveContainer" containerID="1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e" Nov 25 22:06:42 crc kubenswrapper[4931]: E1125 22:06:42.653979 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e\": container with ID starting with 1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e not found: ID does not exist" containerID="1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.654009 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e"} err="failed to get container status \"1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e\": rpc error: code = NotFound desc = could not find container \"1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e\": container with ID starting with 1e966b43b968903b85d9391997583d3ca32c666f2fbed2efc8c7b5f5089c758e not found: ID does not exist" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.654030 4931 scope.go:117] "RemoveContainer" containerID="d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3" Nov 25 22:06:42 crc kubenswrapper[4931]: E1125 22:06:42.654330 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3\": container with ID starting with d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3 not found: ID does not exist" containerID="d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.654355 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3"} err="failed to get container status \"d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3\": rpc error: code = NotFound desc = could not find container \"d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3\": container with ID starting with d969b9b8ba03eb31714a1c4ca0efd32070a507e61b32661c64f093a47da18ba3 not found: ID does not exist" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.654370 4931 scope.go:117] "RemoveContainer" containerID="5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6" Nov 25 22:06:42 crc kubenswrapper[4931]: E1125 22:06:42.654592 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6\": container with ID starting with 5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6 not found: ID does not exist" containerID="5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.654609 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6"} err="failed to get container status \"5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6\": rpc error: code = NotFound desc = could not find container \"5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6\": container with ID starting with 5903a115ba7ada821858ed1a0a4c61878c61af4c1653c9c685a691084538f7d6 not found: ID does not exist" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691069 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98"] Nov 25 22:06:42 crc kubenswrapper[4931]: E1125 22:06:42.691308 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerName="extract-utilities" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691328 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerName="extract-utilities" Nov 25 22:06:42 crc kubenswrapper[4931]: E1125 22:06:42.691342 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerName="registry-server" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691350 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerName="registry-server" Nov 25 22:06:42 crc kubenswrapper[4931]: E1125 22:06:42.691364 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b979640a-23b5-48df-96be-82bc4630d172" containerName="extract" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691373 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b979640a-23b5-48df-96be-82bc4630d172" containerName="extract" Nov 25 22:06:42 crc kubenswrapper[4931]: E1125 22:06:42.691388 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b979640a-23b5-48df-96be-82bc4630d172" containerName="pull" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691395 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b979640a-23b5-48df-96be-82bc4630d172" containerName="pull" Nov 25 22:06:42 crc kubenswrapper[4931]: E1125 22:06:42.691410 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b979640a-23b5-48df-96be-82bc4630d172" containerName="util" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691415 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b979640a-23b5-48df-96be-82bc4630d172" containerName="util" Nov 25 22:06:42 crc kubenswrapper[4931]: E1125 22:06:42.691424 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerName="extract-content" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691430 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerName="extract-content" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691534 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b979640a-23b5-48df-96be-82bc4630d172" containerName="extract" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691548 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" containerName="registry-server" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.691941 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.694160 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-5sjd4" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.716081 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98"] Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.771019 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99gm4\" (UniqueName: \"kubernetes.io/projected/7b1bb711-ce82-4ea4-9667-ff4816b162f4-kube-api-access-99gm4\") pod \"rabbitmq-cluster-operator-779fc9694b-zsd98\" (UID: \"7b1bb711-ce82-4ea4-9667-ff4816b162f4\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.852384 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3f4937e-64ff-4f9a-addd-0d1f436252d1" (UID: "a3f4937e-64ff-4f9a-addd-0d1f436252d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.872354 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99gm4\" (UniqueName: \"kubernetes.io/projected/7b1bb711-ce82-4ea4-9667-ff4816b162f4-kube-api-access-99gm4\") pod \"rabbitmq-cluster-operator-779fc9694b-zsd98\" (UID: \"7b1bb711-ce82-4ea4-9667-ff4816b162f4\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.872442 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f4937e-64ff-4f9a-addd-0d1f436252d1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.891009 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99gm4\" (UniqueName: \"kubernetes.io/projected/7b1bb711-ce82-4ea4-9667-ff4816b162f4-kube-api-access-99gm4\") pod \"rabbitmq-cluster-operator-779fc9694b-zsd98\" (UID: \"7b1bb711-ce82-4ea4-9667-ff4816b162f4\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.929998 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tglrb"] Nov 25 22:06:42 crc kubenswrapper[4931]: I1125 22:06:42.934370 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tglrb"] Nov 25 22:06:43 crc kubenswrapper[4931]: I1125 22:06:43.007994 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" Nov 25 22:06:43 crc kubenswrapper[4931]: I1125 22:06:43.433450 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98"] Nov 25 22:06:43 crc kubenswrapper[4931]: I1125 22:06:43.590993 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" event={"ID":"7b1bb711-ce82-4ea4-9667-ff4816b162f4","Type":"ContainerStarted","Data":"e543984a109d84a50e6b45f34c0ffc33a1c47d89330528aca0b4c2c8d7226768"} Nov 25 22:06:44 crc kubenswrapper[4931]: I1125 22:06:44.430417 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3f4937e-64ff-4f9a-addd-0d1f436252d1" path="/var/lib/kubelet/pods/a3f4937e-64ff-4f9a-addd-0d1f436252d1/volumes" Nov 25 22:06:47 crc kubenswrapper[4931]: I1125 22:06:47.625534 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" event={"ID":"7b1bb711-ce82-4ea4-9667-ff4816b162f4","Type":"ContainerStarted","Data":"6f212b00aeed7232535f9c5ae6ef70d445c379d0b5f19f7dd4fdbe485d5236b1"} Nov 25 22:06:47 crc kubenswrapper[4931]: I1125 22:06:47.641941 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" podStartSLOduration=2.382643851 podStartE2EDuration="5.641923669s" podCreationTimestamp="2025-11-25 22:06:42 +0000 UTC" firstStartedPulling="2025-11-25 22:06:43.442865178 +0000 UTC m=+941.799377297" lastFinishedPulling="2025-11-25 22:06:46.702145006 +0000 UTC m=+945.058657115" observedRunningTime="2025-11-25 22:06:47.640234783 +0000 UTC m=+945.996746922" watchObservedRunningTime="2025-11-25 22:06:47.641923669 +0000 UTC m=+945.998435788" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.921676 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.923056 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.935640 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.935663 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.935743 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.935931 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-b9jg9" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.935649 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.975670 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.999775 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.999825 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.999853 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rffr\" (UniqueName: \"kubernetes.io/projected/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-kube-api-access-4rffr\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.999909 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:50 crc kubenswrapper[4931]: I1125 22:06:50.999951 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ab57b318-6fd5-4957-bd13-e812d5cbc301\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab57b318-6fd5-4957-bd13-e812d5cbc301\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.000069 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.000137 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.000210 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.101390 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.101445 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.101494 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.101525 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.101552 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rffr\" (UniqueName: \"kubernetes.io/projected/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-kube-api-access-4rffr\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.101578 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.101616 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ab57b318-6fd5-4957-bd13-e812d5cbc301\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab57b318-6fd5-4957-bd13-e812d5cbc301\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.101648 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.102163 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.102805 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.103844 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.109227 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.109937 4931 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.109968 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ab57b318-6fd5-4957-bd13-e812d5cbc301\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab57b318-6fd5-4957-bd13-e812d5cbc301\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5414c56b7b401f17d9f85ba111a9a059b05dc7bfc139177229da0f4ab034d0bb/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.110006 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.113141 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.132496 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rffr\" (UniqueName: \"kubernetes.io/projected/4b3d80c3-c790-47e0-ae6c-5c639da2ef84-kube-api-access-4rffr\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.148756 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ab57b318-6fd5-4957-bd13-e812d5cbc301\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab57b318-6fd5-4957-bd13-e812d5cbc301\") pod \"rabbitmq-server-0\" (UID: \"4b3d80c3-c790-47e0-ae6c-5c639da2ef84\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.262372 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.564093 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 22:06:51 crc kubenswrapper[4931]: W1125 22:06:51.568823 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b3d80c3_c790_47e0_ae6c_5c639da2ef84.slice/crio-8fd08232836eb62f74bc64622b74928325dc0dd0366d529926886f5783aeb71e WatchSource:0}: Error finding container 8fd08232836eb62f74bc64622b74928325dc0dd0366d529926886f5783aeb71e: Status 404 returned error can't find the container with id 8fd08232836eb62f74bc64622b74928325dc0dd0366d529926886f5783aeb71e Nov 25 22:06:51 crc kubenswrapper[4931]: I1125 22:06:51.652792 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"4b3d80c3-c790-47e0-ae6c-5c639da2ef84","Type":"ContainerStarted","Data":"8fd08232836eb62f74bc64622b74928325dc0dd0366d529926886f5783aeb71e"} Nov 25 22:06:52 crc kubenswrapper[4931]: I1125 22:06:52.300699 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-lt5k5"] Nov 25 22:06:52 crc kubenswrapper[4931]: I1125 22:06:52.302348 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-lt5k5" Nov 25 22:06:52 crc kubenswrapper[4931]: I1125 22:06:52.306455 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-jlksk" Nov 25 22:06:52 crc kubenswrapper[4931]: I1125 22:06:52.324273 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-lt5k5"] Nov 25 22:06:52 crc kubenswrapper[4931]: I1125 22:06:52.424779 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tpnm\" (UniqueName: \"kubernetes.io/projected/b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e-kube-api-access-8tpnm\") pod \"keystone-operator-index-lt5k5\" (UID: \"b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e\") " pod="openstack-operators/keystone-operator-index-lt5k5" Nov 25 22:06:52 crc kubenswrapper[4931]: I1125 22:06:52.526220 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tpnm\" (UniqueName: \"kubernetes.io/projected/b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e-kube-api-access-8tpnm\") pod \"keystone-operator-index-lt5k5\" (UID: \"b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e\") " pod="openstack-operators/keystone-operator-index-lt5k5" Nov 25 22:06:52 crc kubenswrapper[4931]: I1125 22:06:52.554591 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tpnm\" (UniqueName: \"kubernetes.io/projected/b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e-kube-api-access-8tpnm\") pod \"keystone-operator-index-lt5k5\" (UID: \"b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e\") " pod="openstack-operators/keystone-operator-index-lt5k5" Nov 25 22:06:52 crc kubenswrapper[4931]: I1125 22:06:52.655166 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-lt5k5" Nov 25 22:06:52 crc kubenswrapper[4931]: I1125 22:06:52.873602 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-lt5k5"] Nov 25 22:06:53 crc kubenswrapper[4931]: I1125 22:06:53.668202 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-lt5k5" event={"ID":"b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e","Type":"ContainerStarted","Data":"5aac531e26c06a544eca4cae536cdd5981a0e2ff44399703d6581c2f1ce64520"} Nov 25 22:06:54 crc kubenswrapper[4931]: I1125 22:06:54.688752 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-lt5k5" event={"ID":"b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e","Type":"ContainerStarted","Data":"15714bd047b56f5019c92519635806fbbb9ff849323b6c824b910baccba66657"} Nov 25 22:06:56 crc kubenswrapper[4931]: I1125 22:06:56.700110 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-lt5k5" podStartSLOduration=3.90251982 podStartE2EDuration="4.700088136s" podCreationTimestamp="2025-11-25 22:06:52 +0000 UTC" firstStartedPulling="2025-11-25 22:06:52.904321049 +0000 UTC m=+951.260833158" lastFinishedPulling="2025-11-25 22:06:53.701889355 +0000 UTC m=+952.058401474" observedRunningTime="2025-11-25 22:06:54.711486326 +0000 UTC m=+953.067998495" watchObservedRunningTime="2025-11-25 22:06:56.700088136 +0000 UTC m=+955.056600245" Nov 25 22:06:56 crc kubenswrapper[4931]: I1125 22:06:56.700537 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-lt5k5"] Nov 25 22:06:56 crc kubenswrapper[4931]: I1125 22:06:56.704381 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-lt5k5" podUID="b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e" containerName="registry-server" containerID="cri-o://15714bd047b56f5019c92519635806fbbb9ff849323b6c824b910baccba66657" gracePeriod=2 Nov 25 22:06:57 crc kubenswrapper[4931]: I1125 22:06:57.309301 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-8gl8x"] Nov 25 22:06:57 crc kubenswrapper[4931]: I1125 22:06:57.311019 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-8gl8x" Nov 25 22:06:57 crc kubenswrapper[4931]: I1125 22:06:57.317235 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-8gl8x"] Nov 25 22:06:57 crc kubenswrapper[4931]: I1125 22:06:57.409191 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb4lb\" (UniqueName: \"kubernetes.io/projected/d83baa8b-fccc-4dd8-9fd9-f7fe5850e5ca-kube-api-access-xb4lb\") pod \"keystone-operator-index-8gl8x\" (UID: \"d83baa8b-fccc-4dd8-9fd9-f7fe5850e5ca\") " pod="openstack-operators/keystone-operator-index-8gl8x" Nov 25 22:06:57 crc kubenswrapper[4931]: I1125 22:06:57.510545 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb4lb\" (UniqueName: \"kubernetes.io/projected/d83baa8b-fccc-4dd8-9fd9-f7fe5850e5ca-kube-api-access-xb4lb\") pod \"keystone-operator-index-8gl8x\" (UID: \"d83baa8b-fccc-4dd8-9fd9-f7fe5850e5ca\") " pod="openstack-operators/keystone-operator-index-8gl8x" Nov 25 22:06:57 crc kubenswrapper[4931]: I1125 22:06:57.551617 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb4lb\" (UniqueName: \"kubernetes.io/projected/d83baa8b-fccc-4dd8-9fd9-f7fe5850e5ca-kube-api-access-xb4lb\") pod \"keystone-operator-index-8gl8x\" (UID: \"d83baa8b-fccc-4dd8-9fd9-f7fe5850e5ca\") " pod="openstack-operators/keystone-operator-index-8gl8x" Nov 25 22:06:57 crc kubenswrapper[4931]: I1125 22:06:57.638375 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-8gl8x" Nov 25 22:06:57 crc kubenswrapper[4931]: I1125 22:06:57.744394 4931 generic.go:334] "Generic (PLEG): container finished" podID="b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e" containerID="15714bd047b56f5019c92519635806fbbb9ff849323b6c824b910baccba66657" exitCode=0 Nov 25 22:06:57 crc kubenswrapper[4931]: I1125 22:06:57.744714 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-lt5k5" event={"ID":"b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e","Type":"ContainerDied","Data":"15714bd047b56f5019c92519635806fbbb9ff849323b6c824b910baccba66657"} Nov 25 22:06:58 crc kubenswrapper[4931]: I1125 22:06:58.535160 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-lt5k5" Nov 25 22:06:58 crc kubenswrapper[4931]: I1125 22:06:58.626543 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tpnm\" (UniqueName: \"kubernetes.io/projected/b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e-kube-api-access-8tpnm\") pod \"b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e\" (UID: \"b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e\") " Nov 25 22:06:58 crc kubenswrapper[4931]: I1125 22:06:58.635515 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e-kube-api-access-8tpnm" (OuterVolumeSpecName: "kube-api-access-8tpnm") pod "b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e" (UID: "b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e"). InnerVolumeSpecName "kube-api-access-8tpnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:06:58 crc kubenswrapper[4931]: I1125 22:06:58.728396 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tpnm\" (UniqueName: \"kubernetes.io/projected/b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e-kube-api-access-8tpnm\") on node \"crc\" DevicePath \"\"" Nov 25 22:06:58 crc kubenswrapper[4931]: I1125 22:06:58.752790 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-lt5k5" event={"ID":"b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e","Type":"ContainerDied","Data":"5aac531e26c06a544eca4cae536cdd5981a0e2ff44399703d6581c2f1ce64520"} Nov 25 22:06:58 crc kubenswrapper[4931]: I1125 22:06:58.752844 4931 scope.go:117] "RemoveContainer" containerID="15714bd047b56f5019c92519635806fbbb9ff849323b6c824b910baccba66657" Nov 25 22:06:58 crc kubenswrapper[4931]: I1125 22:06:58.753946 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-lt5k5" Nov 25 22:06:58 crc kubenswrapper[4931]: I1125 22:06:58.781041 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-lt5k5"] Nov 25 22:06:58 crc kubenswrapper[4931]: I1125 22:06:58.784950 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-lt5k5"] Nov 25 22:06:59 crc kubenswrapper[4931]: I1125 22:06:59.137237 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-8gl8x"] Nov 25 22:06:59 crc kubenswrapper[4931]: I1125 22:06:59.763955 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-8gl8x" event={"ID":"d83baa8b-fccc-4dd8-9fd9-f7fe5850e5ca","Type":"ContainerStarted","Data":"6e8f32f914b37af4e81958be28acfc4fb6a2c20dbe89ad4fb84df2dc5c6d904b"} Nov 25 22:07:00 crc kubenswrapper[4931]: I1125 22:07:00.438316 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e" path="/var/lib/kubelet/pods/b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e/volumes" Nov 25 22:07:00 crc kubenswrapper[4931]: I1125 22:07:00.776248 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-8gl8x" event={"ID":"d83baa8b-fccc-4dd8-9fd9-f7fe5850e5ca","Type":"ContainerStarted","Data":"4e25ae37e1ea98fe2606b0718677e3af89b5d1ce7dd366c0e4e86580c927ef30"} Nov 25 22:07:00 crc kubenswrapper[4931]: I1125 22:07:00.811309 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-8gl8x" podStartSLOduration=3.348641302 podStartE2EDuration="3.811279163s" podCreationTimestamp="2025-11-25 22:06:57 +0000 UTC" firstStartedPulling="2025-11-25 22:06:59.304563782 +0000 UTC m=+957.661075911" lastFinishedPulling="2025-11-25 22:06:59.767201613 +0000 UTC m=+958.123713772" observedRunningTime="2025-11-25 22:07:00.802753834 +0000 UTC m=+959.159266003" watchObservedRunningTime="2025-11-25 22:07:00.811279163 +0000 UTC m=+959.167791312" Nov 25 22:07:01 crc kubenswrapper[4931]: I1125 22:07:01.786689 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"4b3d80c3-c790-47e0-ae6c-5c639da2ef84","Type":"ContainerStarted","Data":"a8f18511250feaa41cd0f717bd2114ff900ec680179640275261085abfe8f14e"} Nov 25 22:07:07 crc kubenswrapper[4931]: I1125 22:07:07.641620 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-8gl8x" Nov 25 22:07:07 crc kubenswrapper[4931]: I1125 22:07:07.642978 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-8gl8x" Nov 25 22:07:07 crc kubenswrapper[4931]: I1125 22:07:07.675501 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-8gl8x" Nov 25 22:07:07 crc kubenswrapper[4931]: I1125 22:07:07.878575 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-8gl8x" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.753244 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24"] Nov 25 22:07:10 crc kubenswrapper[4931]: E1125 22:07:10.754172 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e" containerName="registry-server" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.754202 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e" containerName="registry-server" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.754485 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6b4f7ad-141b-4dfb-9cf7-619e1c04aa6e" containerName="registry-server" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.756464 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.759358 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrsb7" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.791967 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24"] Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.839649 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d974k\" (UniqueName: \"kubernetes.io/projected/3039d342-7a69-4ad3-bd7e-e21a65264f61-kube-api-access-d974k\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.839850 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.839980 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.941592 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d974k\" (UniqueName: \"kubernetes.io/projected/3039d342-7a69-4ad3-bd7e-e21a65264f61-kube-api-access-d974k\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.941641 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.941683 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.942273 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.942973 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:10 crc kubenswrapper[4931]: I1125 22:07:10.982381 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d974k\" (UniqueName: \"kubernetes.io/projected/3039d342-7a69-4ad3-bd7e-e21a65264f61-kube-api-access-d974k\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:11 crc kubenswrapper[4931]: I1125 22:07:11.097190 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:11 crc kubenswrapper[4931]: I1125 22:07:11.557687 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24"] Nov 25 22:07:11 crc kubenswrapper[4931]: I1125 22:07:11.896471 4931 generic.go:334] "Generic (PLEG): container finished" podID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerID="b72979844ca4d266f71e4c541379fc9252e512afc5389adb5b34f0f14f98e7cd" exitCode=0 Nov 25 22:07:11 crc kubenswrapper[4931]: I1125 22:07:11.896537 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" event={"ID":"3039d342-7a69-4ad3-bd7e-e21a65264f61","Type":"ContainerDied","Data":"b72979844ca4d266f71e4c541379fc9252e512afc5389adb5b34f0f14f98e7cd"} Nov 25 22:07:11 crc kubenswrapper[4931]: I1125 22:07:11.896588 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" event={"ID":"3039d342-7a69-4ad3-bd7e-e21a65264f61","Type":"ContainerStarted","Data":"33a8b7aaa3dc8ae8079219f473762ec53e15073a75963eba6ea60e6025058453"} Nov 25 22:07:12 crc kubenswrapper[4931]: I1125 22:07:12.903576 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" event={"ID":"3039d342-7a69-4ad3-bd7e-e21a65264f61","Type":"ContainerStarted","Data":"894c09fe799c084d82310fe31a20faa80e9f9bed0fdce95c11192d5e4202cf7f"} Nov 25 22:07:13 crc kubenswrapper[4931]: I1125 22:07:13.915214 4931 generic.go:334] "Generic (PLEG): container finished" podID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerID="894c09fe799c084d82310fe31a20faa80e9f9bed0fdce95c11192d5e4202cf7f" exitCode=0 Nov 25 22:07:13 crc kubenswrapper[4931]: I1125 22:07:13.915282 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" event={"ID":"3039d342-7a69-4ad3-bd7e-e21a65264f61","Type":"ContainerDied","Data":"894c09fe799c084d82310fe31a20faa80e9f9bed0fdce95c11192d5e4202cf7f"} Nov 25 22:07:14 crc kubenswrapper[4931]: I1125 22:07:14.925646 4931 generic.go:334] "Generic (PLEG): container finished" podID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerID="87f0285acc185ada5bca15ce76aa8b8919affa337b345290fe0a83985c66b7e8" exitCode=0 Nov 25 22:07:14 crc kubenswrapper[4931]: I1125 22:07:14.925684 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" event={"ID":"3039d342-7a69-4ad3-bd7e-e21a65264f61","Type":"ContainerDied","Data":"87f0285acc185ada5bca15ce76aa8b8919affa337b345290fe0a83985c66b7e8"} Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.235043 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.326730 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-util\") pod \"3039d342-7a69-4ad3-bd7e-e21a65264f61\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.326786 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d974k\" (UniqueName: \"kubernetes.io/projected/3039d342-7a69-4ad3-bd7e-e21a65264f61-kube-api-access-d974k\") pod \"3039d342-7a69-4ad3-bd7e-e21a65264f61\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.326838 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-bundle\") pod \"3039d342-7a69-4ad3-bd7e-e21a65264f61\" (UID: \"3039d342-7a69-4ad3-bd7e-e21a65264f61\") " Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.328503 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-bundle" (OuterVolumeSpecName: "bundle") pod "3039d342-7a69-4ad3-bd7e-e21a65264f61" (UID: "3039d342-7a69-4ad3-bd7e-e21a65264f61"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.333959 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3039d342-7a69-4ad3-bd7e-e21a65264f61-kube-api-access-d974k" (OuterVolumeSpecName: "kube-api-access-d974k") pod "3039d342-7a69-4ad3-bd7e-e21a65264f61" (UID: "3039d342-7a69-4ad3-bd7e-e21a65264f61"). InnerVolumeSpecName "kube-api-access-d974k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.357807 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-util" (OuterVolumeSpecName: "util") pod "3039d342-7a69-4ad3-bd7e-e21a65264f61" (UID: "3039d342-7a69-4ad3-bd7e-e21a65264f61"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.428263 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-util\") on node \"crc\" DevicePath \"\"" Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.428300 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d974k\" (UniqueName: \"kubernetes.io/projected/3039d342-7a69-4ad3-bd7e-e21a65264f61-kube-api-access-d974k\") on node \"crc\" DevicePath \"\"" Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.428315 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3039d342-7a69-4ad3-bd7e-e21a65264f61-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.945318 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" event={"ID":"3039d342-7a69-4ad3-bd7e-e21a65264f61","Type":"ContainerDied","Data":"33a8b7aaa3dc8ae8079219f473762ec53e15073a75963eba6ea60e6025058453"} Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.945387 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33a8b7aaa3dc8ae8079219f473762ec53e15073a75963eba6ea60e6025058453" Nov 25 22:07:16 crc kubenswrapper[4931]: I1125 22:07:16.945420 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24" Nov 25 22:07:26 crc kubenswrapper[4931]: I1125 22:07:26.975562 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc"] Nov 25 22:07:26 crc kubenswrapper[4931]: E1125 22:07:26.976276 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerName="util" Nov 25 22:07:26 crc kubenswrapper[4931]: I1125 22:07:26.976291 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerName="util" Nov 25 22:07:26 crc kubenswrapper[4931]: E1125 22:07:26.976309 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerName="pull" Nov 25 22:07:26 crc kubenswrapper[4931]: I1125 22:07:26.976317 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerName="pull" Nov 25 22:07:26 crc kubenswrapper[4931]: E1125 22:07:26.976336 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerName="extract" Nov 25 22:07:26 crc kubenswrapper[4931]: I1125 22:07:26.976345 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerName="extract" Nov 25 22:07:26 crc kubenswrapper[4931]: I1125 22:07:26.976479 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="3039d342-7a69-4ad3-bd7e-e21a65264f61" containerName="extract" Nov 25 22:07:26 crc kubenswrapper[4931]: I1125 22:07:26.976980 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:26 crc kubenswrapper[4931]: I1125 22:07:26.979435 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 25 22:07:26 crc kubenswrapper[4931]: I1125 22:07:26.979648 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vd8m5" Nov 25 22:07:26 crc kubenswrapper[4931]: I1125 22:07:26.986611 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc"] Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.078832 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d9wn\" (UniqueName: \"kubernetes.io/projected/044c2639-0ceb-4be3-8326-4fdc552aefb6-kube-api-access-4d9wn\") pod \"keystone-operator-controller-manager-5b59697cf5-jthhc\" (UID: \"044c2639-0ceb-4be3-8326-4fdc552aefb6\") " pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.078994 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/044c2639-0ceb-4be3-8326-4fdc552aefb6-apiservice-cert\") pod \"keystone-operator-controller-manager-5b59697cf5-jthhc\" (UID: \"044c2639-0ceb-4be3-8326-4fdc552aefb6\") " pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.079081 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/044c2639-0ceb-4be3-8326-4fdc552aefb6-webhook-cert\") pod \"keystone-operator-controller-manager-5b59697cf5-jthhc\" (UID: \"044c2639-0ceb-4be3-8326-4fdc552aefb6\") " pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.180277 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d9wn\" (UniqueName: \"kubernetes.io/projected/044c2639-0ceb-4be3-8326-4fdc552aefb6-kube-api-access-4d9wn\") pod \"keystone-operator-controller-manager-5b59697cf5-jthhc\" (UID: \"044c2639-0ceb-4be3-8326-4fdc552aefb6\") " pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.180384 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/044c2639-0ceb-4be3-8326-4fdc552aefb6-apiservice-cert\") pod \"keystone-operator-controller-manager-5b59697cf5-jthhc\" (UID: \"044c2639-0ceb-4be3-8326-4fdc552aefb6\") " pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.180471 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/044c2639-0ceb-4be3-8326-4fdc552aefb6-webhook-cert\") pod \"keystone-operator-controller-manager-5b59697cf5-jthhc\" (UID: \"044c2639-0ceb-4be3-8326-4fdc552aefb6\") " pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.187714 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/044c2639-0ceb-4be3-8326-4fdc552aefb6-apiservice-cert\") pod \"keystone-operator-controller-manager-5b59697cf5-jthhc\" (UID: \"044c2639-0ceb-4be3-8326-4fdc552aefb6\") " pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.188321 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/044c2639-0ceb-4be3-8326-4fdc552aefb6-webhook-cert\") pod \"keystone-operator-controller-manager-5b59697cf5-jthhc\" (UID: \"044c2639-0ceb-4be3-8326-4fdc552aefb6\") " pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.203589 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d9wn\" (UniqueName: \"kubernetes.io/projected/044c2639-0ceb-4be3-8326-4fdc552aefb6-kube-api-access-4d9wn\") pod \"keystone-operator-controller-manager-5b59697cf5-jthhc\" (UID: \"044c2639-0ceb-4be3-8326-4fdc552aefb6\") " pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.303143 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:27 crc kubenswrapper[4931]: I1125 22:07:27.817361 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc"] Nov 25 22:07:28 crc kubenswrapper[4931]: I1125 22:07:28.024147 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" event={"ID":"044c2639-0ceb-4be3-8326-4fdc552aefb6","Type":"ContainerStarted","Data":"0dd654848a2e428849c2958d8e6b92a0ec14af78beb23a77ae68fc955bf1cdba"} Nov 25 22:07:33 crc kubenswrapper[4931]: I1125 22:07:33.061144 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" event={"ID":"044c2639-0ceb-4be3-8326-4fdc552aefb6","Type":"ContainerStarted","Data":"a51a02fea4208b822ab73a8a98ebe332fbe70871d27bff531648f01d2a486f6f"} Nov 25 22:07:33 crc kubenswrapper[4931]: I1125 22:07:33.061628 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:33 crc kubenswrapper[4931]: I1125 22:07:33.079289 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" podStartSLOduration=2.4630959199999998 podStartE2EDuration="7.079269278s" podCreationTimestamp="2025-11-25 22:07:26 +0000 UTC" firstStartedPulling="2025-11-25 22:07:27.832941681 +0000 UTC m=+986.189453840" lastFinishedPulling="2025-11-25 22:07:32.449115079 +0000 UTC m=+990.805627198" observedRunningTime="2025-11-25 22:07:33.076996058 +0000 UTC m=+991.433508187" watchObservedRunningTime="2025-11-25 22:07:33.079269278 +0000 UTC m=+991.435781397" Nov 25 22:07:34 crc kubenswrapper[4931]: I1125 22:07:34.071151 4931 generic.go:334] "Generic (PLEG): container finished" podID="4b3d80c3-c790-47e0-ae6c-5c639da2ef84" containerID="a8f18511250feaa41cd0f717bd2114ff900ec680179640275261085abfe8f14e" exitCode=0 Nov 25 22:07:34 crc kubenswrapper[4931]: I1125 22:07:34.071252 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"4b3d80c3-c790-47e0-ae6c-5c639da2ef84","Type":"ContainerDied","Data":"a8f18511250feaa41cd0f717bd2114ff900ec680179640275261085abfe8f14e"} Nov 25 22:07:35 crc kubenswrapper[4931]: I1125 22:07:35.082275 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"4b3d80c3-c790-47e0-ae6c-5c639da2ef84","Type":"ContainerStarted","Data":"34d68a84470aedf9f24ea2e2861a5af66260ddbb8bf50be30ed74b92c9e63f78"} Nov 25 22:07:35 crc kubenswrapper[4931]: I1125 22:07:35.082618 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:07:35 crc kubenswrapper[4931]: I1125 22:07:35.121255 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=38.337704862 podStartE2EDuration="46.121228385s" podCreationTimestamp="2025-11-25 22:06:49 +0000 UTC" firstStartedPulling="2025-11-25 22:06:51.572072163 +0000 UTC m=+949.928584292" lastFinishedPulling="2025-11-25 22:06:59.355595696 +0000 UTC m=+957.712107815" observedRunningTime="2025-11-25 22:07:35.113347603 +0000 UTC m=+993.469859752" watchObservedRunningTime="2025-11-25 22:07:35.121228385 +0000 UTC m=+993.477740504" Nov 25 22:07:37 crc kubenswrapper[4931]: I1125 22:07:37.309514 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:07:38 crc kubenswrapper[4931]: I1125 22:07:38.843707 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:07:38 crc kubenswrapper[4931]: I1125 22:07:38.844343 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:07:42 crc kubenswrapper[4931]: I1125 22:07:42.904428 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-mhxw4"] Nov 25 22:07:42 crc kubenswrapper[4931]: I1125 22:07:42.906275 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-mhxw4" Nov 25 22:07:42 crc kubenswrapper[4931]: I1125 22:07:42.909136 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-qsxk5" Nov 25 22:07:42 crc kubenswrapper[4931]: I1125 22:07:42.915939 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-mhxw4"] Nov 25 22:07:43 crc kubenswrapper[4931]: I1125 22:07:43.079764 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wxk9\" (UniqueName: \"kubernetes.io/projected/502b379e-39c4-4a24-a639-7aa82885c7fd-kube-api-access-8wxk9\") pod \"horizon-operator-index-mhxw4\" (UID: \"502b379e-39c4-4a24-a639-7aa82885c7fd\") " pod="openstack-operators/horizon-operator-index-mhxw4" Nov 25 22:07:43 crc kubenswrapper[4931]: I1125 22:07:43.181337 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wxk9\" (UniqueName: \"kubernetes.io/projected/502b379e-39c4-4a24-a639-7aa82885c7fd-kube-api-access-8wxk9\") pod \"horizon-operator-index-mhxw4\" (UID: \"502b379e-39c4-4a24-a639-7aa82885c7fd\") " pod="openstack-operators/horizon-operator-index-mhxw4" Nov 25 22:07:43 crc kubenswrapper[4931]: I1125 22:07:43.201990 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wxk9\" (UniqueName: \"kubernetes.io/projected/502b379e-39c4-4a24-a639-7aa82885c7fd-kube-api-access-8wxk9\") pod \"horizon-operator-index-mhxw4\" (UID: \"502b379e-39c4-4a24-a639-7aa82885c7fd\") " pod="openstack-operators/horizon-operator-index-mhxw4" Nov 25 22:07:43 crc kubenswrapper[4931]: I1125 22:07:43.226789 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-mhxw4" Nov 25 22:07:43 crc kubenswrapper[4931]: I1125 22:07:43.672548 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-mhxw4"] Nov 25 22:07:44 crc kubenswrapper[4931]: I1125 22:07:44.154042 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-mhxw4" event={"ID":"502b379e-39c4-4a24-a639-7aa82885c7fd","Type":"ContainerStarted","Data":"3e38907ef654128e218fd732f4eb47386168c405f9f49de09ece601aa9018432"} Nov 25 22:07:45 crc kubenswrapper[4931]: I1125 22:07:45.495309 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-4v74n"] Nov 25 22:07:45 crc kubenswrapper[4931]: I1125 22:07:45.496687 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-4v74n" Nov 25 22:07:45 crc kubenswrapper[4931]: I1125 22:07:45.498329 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-fbp4q" Nov 25 22:07:45 crc kubenswrapper[4931]: I1125 22:07:45.505407 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-4v74n"] Nov 25 22:07:45 crc kubenswrapper[4931]: I1125 22:07:45.621385 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwtnk\" (UniqueName: \"kubernetes.io/projected/6162ec0d-ae9b-4334-a7d0-bc693ee687d2-kube-api-access-rwtnk\") pod \"swift-operator-index-4v74n\" (UID: \"6162ec0d-ae9b-4334-a7d0-bc693ee687d2\") " pod="openstack-operators/swift-operator-index-4v74n" Nov 25 22:07:45 crc kubenswrapper[4931]: I1125 22:07:45.722242 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwtnk\" (UniqueName: \"kubernetes.io/projected/6162ec0d-ae9b-4334-a7d0-bc693ee687d2-kube-api-access-rwtnk\") pod \"swift-operator-index-4v74n\" (UID: \"6162ec0d-ae9b-4334-a7d0-bc693ee687d2\") " pod="openstack-operators/swift-operator-index-4v74n" Nov 25 22:07:45 crc kubenswrapper[4931]: I1125 22:07:45.742797 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwtnk\" (UniqueName: \"kubernetes.io/projected/6162ec0d-ae9b-4334-a7d0-bc693ee687d2-kube-api-access-rwtnk\") pod \"swift-operator-index-4v74n\" (UID: \"6162ec0d-ae9b-4334-a7d0-bc693ee687d2\") " pod="openstack-operators/swift-operator-index-4v74n" Nov 25 22:07:45 crc kubenswrapper[4931]: I1125 22:07:45.814444 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-4v74n" Nov 25 22:07:46 crc kubenswrapper[4931]: I1125 22:07:46.171330 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-mhxw4" event={"ID":"502b379e-39c4-4a24-a639-7aa82885c7fd","Type":"ContainerStarted","Data":"1fb6811e79565c5807ba941388eb82eae2936fe4b750b436348e6bd3bae6473b"} Nov 25 22:07:46 crc kubenswrapper[4931]: I1125 22:07:46.188769 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-mhxw4" podStartSLOduration=2.868280723 podStartE2EDuration="4.188748392s" podCreationTimestamp="2025-11-25 22:07:42 +0000 UTC" firstStartedPulling="2025-11-25 22:07:43.681621646 +0000 UTC m=+1002.038133805" lastFinishedPulling="2025-11-25 22:07:45.002089345 +0000 UTC m=+1003.358601474" observedRunningTime="2025-11-25 22:07:46.186673716 +0000 UTC m=+1004.543185865" watchObservedRunningTime="2025-11-25 22:07:46.188748392 +0000 UTC m=+1004.545260511" Nov 25 22:07:46 crc kubenswrapper[4931]: I1125 22:07:46.338842 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-4v74n"] Nov 25 22:07:46 crc kubenswrapper[4931]: W1125 22:07:46.351574 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6162ec0d_ae9b_4334_a7d0_bc693ee687d2.slice/crio-696cff4d77c6c39b579773859acd4ce7c4a9ce0e9b558a8f41703532f055191a WatchSource:0}: Error finding container 696cff4d77c6c39b579773859acd4ce7c4a9ce0e9b558a8f41703532f055191a: Status 404 returned error can't find the container with id 696cff4d77c6c39b579773859acd4ce7c4a9ce0e9b558a8f41703532f055191a Nov 25 22:07:47 crc kubenswrapper[4931]: I1125 22:07:47.177916 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-4v74n" event={"ID":"6162ec0d-ae9b-4334-a7d0-bc693ee687d2","Type":"ContainerStarted","Data":"696cff4d77c6c39b579773859acd4ce7c4a9ce0e9b558a8f41703532f055191a"} Nov 25 22:07:48 crc kubenswrapper[4931]: I1125 22:07:48.191282 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-4v74n" event={"ID":"6162ec0d-ae9b-4334-a7d0-bc693ee687d2","Type":"ContainerStarted","Data":"261bcb9b8422cc26dd18d5c899e68a1a5b7e9455e6258b8b8b9d2465a6ac27fc"} Nov 25 22:07:48 crc kubenswrapper[4931]: I1125 22:07:48.214655 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-4v74n" podStartSLOduration=2.400385061 podStartE2EDuration="3.214630875s" podCreationTimestamp="2025-11-25 22:07:45 +0000 UTC" firstStartedPulling="2025-11-25 22:07:46.354137543 +0000 UTC m=+1004.710649702" lastFinishedPulling="2025-11-25 22:07:47.168383357 +0000 UTC m=+1005.524895516" observedRunningTime="2025-11-25 22:07:48.207507154 +0000 UTC m=+1006.564019313" watchObservedRunningTime="2025-11-25 22:07:48.214630875 +0000 UTC m=+1006.571143034" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.471021 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-wntmn"] Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.472231 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.474539 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-3765-account-create-update-bvwnb"] Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.475080 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.478007 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.494071 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-3765-account-create-update-bvwnb"] Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.511849 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-wntmn"] Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.597295 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6933bb9-7603-4a25-bc47-685a2e75ce2b-operator-scripts\") pod \"keystone-3765-account-create-update-bvwnb\" (UID: \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\") " pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.597386 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2v2c\" (UniqueName: \"kubernetes.io/projected/4ef44b19-30a3-4d00-ae70-78dc171bc820-kube-api-access-p2v2c\") pod \"keystone-db-create-wntmn\" (UID: \"4ef44b19-30a3-4d00-ae70-78dc171bc820\") " pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.597453 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgqdf\" (UniqueName: \"kubernetes.io/projected/a6933bb9-7603-4a25-bc47-685a2e75ce2b-kube-api-access-cgqdf\") pod \"keystone-3765-account-create-update-bvwnb\" (UID: \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\") " pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.597597 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ef44b19-30a3-4d00-ae70-78dc171bc820-operator-scripts\") pod \"keystone-db-create-wntmn\" (UID: \"4ef44b19-30a3-4d00-ae70-78dc171bc820\") " pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.699112 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6933bb9-7603-4a25-bc47-685a2e75ce2b-operator-scripts\") pod \"keystone-3765-account-create-update-bvwnb\" (UID: \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\") " pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.699169 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2v2c\" (UniqueName: \"kubernetes.io/projected/4ef44b19-30a3-4d00-ae70-78dc171bc820-kube-api-access-p2v2c\") pod \"keystone-db-create-wntmn\" (UID: \"4ef44b19-30a3-4d00-ae70-78dc171bc820\") " pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.699203 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgqdf\" (UniqueName: \"kubernetes.io/projected/a6933bb9-7603-4a25-bc47-685a2e75ce2b-kube-api-access-cgqdf\") pod \"keystone-3765-account-create-update-bvwnb\" (UID: \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\") " pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.699279 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ef44b19-30a3-4d00-ae70-78dc171bc820-operator-scripts\") pod \"keystone-db-create-wntmn\" (UID: \"4ef44b19-30a3-4d00-ae70-78dc171bc820\") " pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.700260 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ef44b19-30a3-4d00-ae70-78dc171bc820-operator-scripts\") pod \"keystone-db-create-wntmn\" (UID: \"4ef44b19-30a3-4d00-ae70-78dc171bc820\") " pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.700626 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6933bb9-7603-4a25-bc47-685a2e75ce2b-operator-scripts\") pod \"keystone-3765-account-create-update-bvwnb\" (UID: \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\") " pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.722173 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgqdf\" (UniqueName: \"kubernetes.io/projected/a6933bb9-7603-4a25-bc47-685a2e75ce2b-kube-api-access-cgqdf\") pod \"keystone-3765-account-create-update-bvwnb\" (UID: \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\") " pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.724604 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2v2c\" (UniqueName: \"kubernetes.io/projected/4ef44b19-30a3-4d00-ae70-78dc171bc820-kube-api-access-p2v2c\") pod \"keystone-db-create-wntmn\" (UID: \"4ef44b19-30a3-4d00-ae70-78dc171bc820\") " pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.802070 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:50 crc kubenswrapper[4931]: I1125 22:07:50.820709 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:51 crc kubenswrapper[4931]: I1125 22:07:51.120494 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-3765-account-create-update-bvwnb"] Nov 25 22:07:51 crc kubenswrapper[4931]: W1125 22:07:51.124274 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6933bb9_7603_4a25_bc47_685a2e75ce2b.slice/crio-b3f23a5e10fd108eef2ca39ba34200b52ca06c92dffc096e03d5c1a3a57145a3 WatchSource:0}: Error finding container b3f23a5e10fd108eef2ca39ba34200b52ca06c92dffc096e03d5c1a3a57145a3: Status 404 returned error can't find the container with id b3f23a5e10fd108eef2ca39ba34200b52ca06c92dffc096e03d5c1a3a57145a3 Nov 25 22:07:51 crc kubenswrapper[4931]: I1125 22:07:51.214410 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" event={"ID":"a6933bb9-7603-4a25-bc47-685a2e75ce2b","Type":"ContainerStarted","Data":"b3f23a5e10fd108eef2ca39ba34200b52ca06c92dffc096e03d5c1a3a57145a3"} Nov 25 22:07:51 crc kubenswrapper[4931]: I1125 22:07:51.263992 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-wntmn"] Nov 25 22:07:51 crc kubenswrapper[4931]: I1125 22:07:51.267052 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 22:07:51 crc kubenswrapper[4931]: W1125 22:07:51.271684 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ef44b19_30a3_4d00_ae70_78dc171bc820.slice/crio-2a1516bb3eb7ba98040a3fd169ce313f0fd3677c862c6444d62d1f6dd5bece27 WatchSource:0}: Error finding container 2a1516bb3eb7ba98040a3fd169ce313f0fd3677c862c6444d62d1f6dd5bece27: Status 404 returned error can't find the container with id 2a1516bb3eb7ba98040a3fd169ce313f0fd3677c862c6444d62d1f6dd5bece27 Nov 25 22:07:52 crc kubenswrapper[4931]: I1125 22:07:52.223331 4931 generic.go:334] "Generic (PLEG): container finished" podID="a6933bb9-7603-4a25-bc47-685a2e75ce2b" containerID="104ceb20bf6a92e6ce2ee2924f41be0756130e1eb6de72fbb776f7c2989ac896" exitCode=0 Nov 25 22:07:52 crc kubenswrapper[4931]: I1125 22:07:52.223870 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" event={"ID":"a6933bb9-7603-4a25-bc47-685a2e75ce2b","Type":"ContainerDied","Data":"104ceb20bf6a92e6ce2ee2924f41be0756130e1eb6de72fbb776f7c2989ac896"} Nov 25 22:07:52 crc kubenswrapper[4931]: I1125 22:07:52.227758 4931 generic.go:334] "Generic (PLEG): container finished" podID="4ef44b19-30a3-4d00-ae70-78dc171bc820" containerID="1ddc3349c2ae687066755465d87ec9e6bccadcfbc00f4358dd9f937123d2a87f" exitCode=0 Nov 25 22:07:52 crc kubenswrapper[4931]: I1125 22:07:52.227985 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-wntmn" event={"ID":"4ef44b19-30a3-4d00-ae70-78dc171bc820","Type":"ContainerDied","Data":"1ddc3349c2ae687066755465d87ec9e6bccadcfbc00f4358dd9f937123d2a87f"} Nov 25 22:07:52 crc kubenswrapper[4931]: I1125 22:07:52.228046 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-wntmn" event={"ID":"4ef44b19-30a3-4d00-ae70-78dc171bc820","Type":"ContainerStarted","Data":"2a1516bb3eb7ba98040a3fd169ce313f0fd3677c862c6444d62d1f6dd5bece27"} Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.227666 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-mhxw4" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.228509 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-mhxw4" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.277496 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-mhxw4" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.678102 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.687693 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.748575 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ef44b19-30a3-4d00-ae70-78dc171bc820-operator-scripts\") pod \"4ef44b19-30a3-4d00-ae70-78dc171bc820\" (UID: \"4ef44b19-30a3-4d00-ae70-78dc171bc820\") " Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.749098 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6933bb9-7603-4a25-bc47-685a2e75ce2b-operator-scripts\") pod \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\" (UID: \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\") " Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.749132 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgqdf\" (UniqueName: \"kubernetes.io/projected/a6933bb9-7603-4a25-bc47-685a2e75ce2b-kube-api-access-cgqdf\") pod \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\" (UID: \"a6933bb9-7603-4a25-bc47-685a2e75ce2b\") " Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.749166 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2v2c\" (UniqueName: \"kubernetes.io/projected/4ef44b19-30a3-4d00-ae70-78dc171bc820-kube-api-access-p2v2c\") pod \"4ef44b19-30a3-4d00-ae70-78dc171bc820\" (UID: \"4ef44b19-30a3-4d00-ae70-78dc171bc820\") " Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.749587 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ef44b19-30a3-4d00-ae70-78dc171bc820-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ef44b19-30a3-4d00-ae70-78dc171bc820" (UID: "4ef44b19-30a3-4d00-ae70-78dc171bc820"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.749690 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6933bb9-7603-4a25-bc47-685a2e75ce2b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6933bb9-7603-4a25-bc47-685a2e75ce2b" (UID: "a6933bb9-7603-4a25-bc47-685a2e75ce2b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.756045 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ef44b19-30a3-4d00-ae70-78dc171bc820-kube-api-access-p2v2c" (OuterVolumeSpecName: "kube-api-access-p2v2c") pod "4ef44b19-30a3-4d00-ae70-78dc171bc820" (UID: "4ef44b19-30a3-4d00-ae70-78dc171bc820"). InnerVolumeSpecName "kube-api-access-p2v2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.756506 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6933bb9-7603-4a25-bc47-685a2e75ce2b-kube-api-access-cgqdf" (OuterVolumeSpecName: "kube-api-access-cgqdf") pod "a6933bb9-7603-4a25-bc47-685a2e75ce2b" (UID: "a6933bb9-7603-4a25-bc47-685a2e75ce2b"). InnerVolumeSpecName "kube-api-access-cgqdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.851244 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgqdf\" (UniqueName: \"kubernetes.io/projected/a6933bb9-7603-4a25-bc47-685a2e75ce2b-kube-api-access-cgqdf\") on node \"crc\" DevicePath \"\"" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.851285 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2v2c\" (UniqueName: \"kubernetes.io/projected/4ef44b19-30a3-4d00-ae70-78dc171bc820-kube-api-access-p2v2c\") on node \"crc\" DevicePath \"\"" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.851301 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ef44b19-30a3-4d00-ae70-78dc171bc820-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:07:53 crc kubenswrapper[4931]: I1125 22:07:53.851314 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6933bb9-7603-4a25-bc47-685a2e75ce2b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:07:54 crc kubenswrapper[4931]: I1125 22:07:54.248355 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" event={"ID":"a6933bb9-7603-4a25-bc47-685a2e75ce2b","Type":"ContainerDied","Data":"b3f23a5e10fd108eef2ca39ba34200b52ca06c92dffc096e03d5c1a3a57145a3"} Nov 25 22:07:54 crc kubenswrapper[4931]: I1125 22:07:54.248390 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3f23a5e10fd108eef2ca39ba34200b52ca06c92dffc096e03d5c1a3a57145a3" Nov 25 22:07:54 crc kubenswrapper[4931]: I1125 22:07:54.248426 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-3765-account-create-update-bvwnb" Nov 25 22:07:54 crc kubenswrapper[4931]: I1125 22:07:54.249717 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-wntmn" Nov 25 22:07:54 crc kubenswrapper[4931]: I1125 22:07:54.249758 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-wntmn" event={"ID":"4ef44b19-30a3-4d00-ae70-78dc171bc820","Type":"ContainerDied","Data":"2a1516bb3eb7ba98040a3fd169ce313f0fd3677c862c6444d62d1f6dd5bece27"} Nov 25 22:07:54 crc kubenswrapper[4931]: I1125 22:07:54.249920 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a1516bb3eb7ba98040a3fd169ce313f0fd3677c862c6444d62d1f6dd5bece27" Nov 25 22:07:54 crc kubenswrapper[4931]: I1125 22:07:54.282477 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-mhxw4" Nov 25 22:07:55 crc kubenswrapper[4931]: I1125 22:07:55.815654 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-4v74n" Nov 25 22:07:55 crc kubenswrapper[4931]: I1125 22:07:55.816060 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-4v74n" Nov 25 22:07:55 crc kubenswrapper[4931]: I1125 22:07:55.848475 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-4v74n" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.030925 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-ngsxr"] Nov 25 22:07:56 crc kubenswrapper[4931]: E1125 22:07:56.031210 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef44b19-30a3-4d00-ae70-78dc171bc820" containerName="mariadb-database-create" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.031237 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef44b19-30a3-4d00-ae70-78dc171bc820" containerName="mariadb-database-create" Nov 25 22:07:56 crc kubenswrapper[4931]: E1125 22:07:56.031261 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6933bb9-7603-4a25-bc47-685a2e75ce2b" containerName="mariadb-account-create-update" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.031274 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6933bb9-7603-4a25-bc47-685a2e75ce2b" containerName="mariadb-account-create-update" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.031454 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6933bb9-7603-4a25-bc47-685a2e75ce2b" containerName="mariadb-account-create-update" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.031470 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ef44b19-30a3-4d00-ae70-78dc171bc820" containerName="mariadb-database-create" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.032152 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.034482 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.034499 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-2j6qp" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.034904 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.034936 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.040448 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-ngsxr"] Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.188224 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575383b0-d5a9-41a6-8799-f7cd7e3e637f-config-data\") pod \"keystone-db-sync-ngsxr\" (UID: \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\") " pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.188369 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jd66\" (UniqueName: \"kubernetes.io/projected/575383b0-d5a9-41a6-8799-f7cd7e3e637f-kube-api-access-9jd66\") pod \"keystone-db-sync-ngsxr\" (UID: \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\") " pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.289735 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jd66\" (UniqueName: \"kubernetes.io/projected/575383b0-d5a9-41a6-8799-f7cd7e3e637f-kube-api-access-9jd66\") pod \"keystone-db-sync-ngsxr\" (UID: \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\") " pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.289912 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575383b0-d5a9-41a6-8799-f7cd7e3e637f-config-data\") pod \"keystone-db-sync-ngsxr\" (UID: \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\") " pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.299545 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575383b0-d5a9-41a6-8799-f7cd7e3e637f-config-data\") pod \"keystone-db-sync-ngsxr\" (UID: \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\") " pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.304701 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-4v74n" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.311328 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jd66\" (UniqueName: \"kubernetes.io/projected/575383b0-d5a9-41a6-8799-f7cd7e3e637f-kube-api-access-9jd66\") pod \"keystone-db-sync-ngsxr\" (UID: \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\") " pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.354824 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:07:56 crc kubenswrapper[4931]: I1125 22:07:56.586452 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-ngsxr"] Nov 25 22:07:57 crc kubenswrapper[4931]: I1125 22:07:57.274687 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-ngsxr" event={"ID":"575383b0-d5a9-41a6-8799-f7cd7e3e637f","Type":"ContainerStarted","Data":"3c1d5ae5f9032b239e015d5acf8f0c24300e7577ddc0f168f7c17804a4667fbe"} Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.326312 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8"] Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.327982 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.330161 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrsb7" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.334079 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8"] Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.389772 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.389835 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqgk5\" (UniqueName: \"kubernetes.io/projected/a748bed2-f75d-4cfb-b20d-3914cb4094ee-kube-api-access-jqgk5\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.389863 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.491085 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.491163 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqgk5\" (UniqueName: \"kubernetes.io/projected/a748bed2-f75d-4cfb-b20d-3914cb4094ee-kube-api-access-jqgk5\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.491196 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.492728 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.493764 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.510729 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqgk5\" (UniqueName: \"kubernetes.io/projected/a748bed2-f75d-4cfb-b20d-3914cb4094ee-kube-api-access-jqgk5\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:02 crc kubenswrapper[4931]: I1125 22:08:02.653921 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.338157 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2"] Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.341405 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.344118 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2"] Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.406437 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqcgg\" (UniqueName: \"kubernetes.io/projected/4237875e-d36d-4c45-9102-2f1242407c46-kube-api-access-gqcgg\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.406700 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.406724 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.508541 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.508723 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqcgg\" (UniqueName: \"kubernetes.io/projected/4237875e-d36d-4c45-9102-2f1242407c46-kube-api-access-gqcgg\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.508764 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.509022 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.509137 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.528522 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqcgg\" (UniqueName: \"kubernetes.io/projected/4237875e-d36d-4c45-9102-2f1242407c46-kube-api-access-gqcgg\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:03 crc kubenswrapper[4931]: I1125 22:08:03.667564 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.029205 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8"] Nov 25 22:08:04 crc kubenswrapper[4931]: W1125 22:08:04.031511 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda748bed2_f75d_4cfb_b20d_3914cb4094ee.slice/crio-92fbaee0eef7980c649e2b8cc023d9ece37f0d380add980a970806d3ccd623d3 WatchSource:0}: Error finding container 92fbaee0eef7980c649e2b8cc023d9ece37f0d380add980a970806d3ccd623d3: Status 404 returned error can't find the container with id 92fbaee0eef7980c649e2b8cc023d9ece37f0d380add980a970806d3ccd623d3 Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.147910 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2"] Nov 25 22:08:04 crc kubenswrapper[4931]: W1125 22:08:04.151562 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4237875e_d36d_4c45_9102_2f1242407c46.slice/crio-263a4f6b7c5891ab65cb6a080692b1f34ccb57b6039a8938979e2179caa9a558 WatchSource:0}: Error finding container 263a4f6b7c5891ab65cb6a080692b1f34ccb57b6039a8938979e2179caa9a558: Status 404 returned error can't find the container with id 263a4f6b7c5891ab65cb6a080692b1f34ccb57b6039a8938979e2179caa9a558 Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.373701 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-ngsxr" event={"ID":"575383b0-d5a9-41a6-8799-f7cd7e3e637f","Type":"ContainerStarted","Data":"97984b2543d5eebb0b54e5e7de0c74158a096a1064e7203da78e19ab861cfe8e"} Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.376407 4931 generic.go:334] "Generic (PLEG): container finished" podID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerID="fe9c74c34fc71094831e4dab380d7a697cfa35acf625a1c897732c6860f5054e" exitCode=0 Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.376635 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" event={"ID":"a748bed2-f75d-4cfb-b20d-3914cb4094ee","Type":"ContainerDied","Data":"fe9c74c34fc71094831e4dab380d7a697cfa35acf625a1c897732c6860f5054e"} Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.376681 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" event={"ID":"a748bed2-f75d-4cfb-b20d-3914cb4094ee","Type":"ContainerStarted","Data":"92fbaee0eef7980c649e2b8cc023d9ece37f0d380add980a970806d3ccd623d3"} Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.378422 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" event={"ID":"4237875e-d36d-4c45-9102-2f1242407c46","Type":"ContainerStarted","Data":"5d4e02a19dd5392f19f233888285f1b888100c73f7bbb3a44a8529e859626d67"} Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.378453 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" event={"ID":"4237875e-d36d-4c45-9102-2f1242407c46","Type":"ContainerStarted","Data":"263a4f6b7c5891ab65cb6a080692b1f34ccb57b6039a8938979e2179caa9a558"} Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.391505 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 22:08:04 crc kubenswrapper[4931]: I1125 22:08:04.414048 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-ngsxr" podStartSLOduration=1.377568392 podStartE2EDuration="8.414025169s" podCreationTimestamp="2025-11-25 22:07:56 +0000 UTC" firstStartedPulling="2025-11-25 22:07:56.591931338 +0000 UTC m=+1014.948443497" lastFinishedPulling="2025-11-25 22:08:03.628388155 +0000 UTC m=+1021.984900274" observedRunningTime="2025-11-25 22:08:04.394486163 +0000 UTC m=+1022.750998312" watchObservedRunningTime="2025-11-25 22:08:04.414025169 +0000 UTC m=+1022.770537288" Nov 25 22:08:05 crc kubenswrapper[4931]: I1125 22:08:05.387588 4931 generic.go:334] "Generic (PLEG): container finished" podID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerID="f1943389f1aeb3b83f3ec7f32183da6dc55156f5c5b97ae09185aae3a6286fda" exitCode=0 Nov 25 22:08:05 crc kubenswrapper[4931]: I1125 22:08:05.387663 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" event={"ID":"a748bed2-f75d-4cfb-b20d-3914cb4094ee","Type":"ContainerDied","Data":"f1943389f1aeb3b83f3ec7f32183da6dc55156f5c5b97ae09185aae3a6286fda"} Nov 25 22:08:05 crc kubenswrapper[4931]: I1125 22:08:05.392748 4931 generic.go:334] "Generic (PLEG): container finished" podID="4237875e-d36d-4c45-9102-2f1242407c46" containerID="5d4e02a19dd5392f19f233888285f1b888100c73f7bbb3a44a8529e859626d67" exitCode=0 Nov 25 22:08:05 crc kubenswrapper[4931]: I1125 22:08:05.392901 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" event={"ID":"4237875e-d36d-4c45-9102-2f1242407c46","Type":"ContainerDied","Data":"5d4e02a19dd5392f19f233888285f1b888100c73f7bbb3a44a8529e859626d67"} Nov 25 22:08:06 crc kubenswrapper[4931]: I1125 22:08:06.405175 4931 generic.go:334] "Generic (PLEG): container finished" podID="4237875e-d36d-4c45-9102-2f1242407c46" containerID="f2387e6e1f98eb5745401acbde5655a0391dcc82adb464da5d61f40c841c186c" exitCode=0 Nov 25 22:08:06 crc kubenswrapper[4931]: I1125 22:08:06.405289 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" event={"ID":"4237875e-d36d-4c45-9102-2f1242407c46","Type":"ContainerDied","Data":"f2387e6e1f98eb5745401acbde5655a0391dcc82adb464da5d61f40c841c186c"} Nov 25 22:08:06 crc kubenswrapper[4931]: I1125 22:08:06.409414 4931 generic.go:334] "Generic (PLEG): container finished" podID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerID="16422804b542c5d97e274a50d167ba9c845179087411dc3150d8fdfa50e7b436" exitCode=0 Nov 25 22:08:06 crc kubenswrapper[4931]: I1125 22:08:06.409450 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" event={"ID":"a748bed2-f75d-4cfb-b20d-3914cb4094ee","Type":"ContainerDied","Data":"16422804b542c5d97e274a50d167ba9c845179087411dc3150d8fdfa50e7b436"} Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.421511 4931 generic.go:334] "Generic (PLEG): container finished" podID="575383b0-d5a9-41a6-8799-f7cd7e3e637f" containerID="97984b2543d5eebb0b54e5e7de0c74158a096a1064e7203da78e19ab861cfe8e" exitCode=0 Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.421589 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-ngsxr" event={"ID":"575383b0-d5a9-41a6-8799-f7cd7e3e637f","Type":"ContainerDied","Data":"97984b2543d5eebb0b54e5e7de0c74158a096a1064e7203da78e19ab861cfe8e"} Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.427491 4931 generic.go:334] "Generic (PLEG): container finished" podID="4237875e-d36d-4c45-9102-2f1242407c46" containerID="732f941d04f2ed3c4caf6db0e9404c52f8d8b2ca45707bf6cb34bc81b452f6a1" exitCode=0 Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.427561 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" event={"ID":"4237875e-d36d-4c45-9102-2f1242407c46","Type":"ContainerDied","Data":"732f941d04f2ed3c4caf6db0e9404c52f8d8b2ca45707bf6cb34bc81b452f6a1"} Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.783212 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.872114 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqgk5\" (UniqueName: \"kubernetes.io/projected/a748bed2-f75d-4cfb-b20d-3914cb4094ee-kube-api-access-jqgk5\") pod \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.872200 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-util\") pod \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.872228 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-bundle\") pod \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\" (UID: \"a748bed2-f75d-4cfb-b20d-3914cb4094ee\") " Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.873140 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-bundle" (OuterVolumeSpecName: "bundle") pod "a748bed2-f75d-4cfb-b20d-3914cb4094ee" (UID: "a748bed2-f75d-4cfb-b20d-3914cb4094ee"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.878296 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a748bed2-f75d-4cfb-b20d-3914cb4094ee-kube-api-access-jqgk5" (OuterVolumeSpecName: "kube-api-access-jqgk5") pod "a748bed2-f75d-4cfb-b20d-3914cb4094ee" (UID: "a748bed2-f75d-4cfb-b20d-3914cb4094ee"). InnerVolumeSpecName "kube-api-access-jqgk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.886224 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-util" (OuterVolumeSpecName: "util") pod "a748bed2-f75d-4cfb-b20d-3914cb4094ee" (UID: "a748bed2-f75d-4cfb-b20d-3914cb4094ee"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.974066 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqgk5\" (UniqueName: \"kubernetes.io/projected/a748bed2-f75d-4cfb-b20d-3914cb4094ee-kube-api-access-jqgk5\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.974120 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-util\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:07 crc kubenswrapper[4931]: I1125 22:08:07.974141 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a748bed2-f75d-4cfb-b20d-3914cb4094ee-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.436094 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.436079 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8" event={"ID":"a748bed2-f75d-4cfb-b20d-3914cb4094ee","Type":"ContainerDied","Data":"92fbaee0eef7980c649e2b8cc023d9ece37f0d380add980a970806d3ccd623d3"} Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.436284 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92fbaee0eef7980c649e2b8cc023d9ece37f0d380add980a970806d3ccd623d3" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.774449 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.778411 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.844010 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.844073 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.887581 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqcgg\" (UniqueName: \"kubernetes.io/projected/4237875e-d36d-4c45-9102-2f1242407c46-kube-api-access-gqcgg\") pod \"4237875e-d36d-4c45-9102-2f1242407c46\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.887681 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575383b0-d5a9-41a6-8799-f7cd7e3e637f-config-data\") pod \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\" (UID: \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\") " Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.887725 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-util\") pod \"4237875e-d36d-4c45-9102-2f1242407c46\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.887812 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jd66\" (UniqueName: \"kubernetes.io/projected/575383b0-d5a9-41a6-8799-f7cd7e3e637f-kube-api-access-9jd66\") pod \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\" (UID: \"575383b0-d5a9-41a6-8799-f7cd7e3e637f\") " Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.887859 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-bundle\") pod \"4237875e-d36d-4c45-9102-2f1242407c46\" (UID: \"4237875e-d36d-4c45-9102-2f1242407c46\") " Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.889413 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-bundle" (OuterVolumeSpecName: "bundle") pod "4237875e-d36d-4c45-9102-2f1242407c46" (UID: "4237875e-d36d-4c45-9102-2f1242407c46"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.891574 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4237875e-d36d-4c45-9102-2f1242407c46-kube-api-access-gqcgg" (OuterVolumeSpecName: "kube-api-access-gqcgg") pod "4237875e-d36d-4c45-9102-2f1242407c46" (UID: "4237875e-d36d-4c45-9102-2f1242407c46"). InnerVolumeSpecName "kube-api-access-gqcgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.896444 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/575383b0-d5a9-41a6-8799-f7cd7e3e637f-kube-api-access-9jd66" (OuterVolumeSpecName: "kube-api-access-9jd66") pod "575383b0-d5a9-41a6-8799-f7cd7e3e637f" (UID: "575383b0-d5a9-41a6-8799-f7cd7e3e637f"). InnerVolumeSpecName "kube-api-access-9jd66". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.902695 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-util" (OuterVolumeSpecName: "util") pod "4237875e-d36d-4c45-9102-2f1242407c46" (UID: "4237875e-d36d-4c45-9102-2f1242407c46"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.929087 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/575383b0-d5a9-41a6-8799-f7cd7e3e637f-config-data" (OuterVolumeSpecName: "config-data") pod "575383b0-d5a9-41a6-8799-f7cd7e3e637f" (UID: "575383b0-d5a9-41a6-8799-f7cd7e3e637f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.989321 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jd66\" (UniqueName: \"kubernetes.io/projected/575383b0-d5a9-41a6-8799-f7cd7e3e637f-kube-api-access-9jd66\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.989805 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.989983 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqcgg\" (UniqueName: \"kubernetes.io/projected/4237875e-d36d-4c45-9102-2f1242407c46-kube-api-access-gqcgg\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.990000 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575383b0-d5a9-41a6-8799-f7cd7e3e637f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:08 crc kubenswrapper[4931]: I1125 22:08:08.990009 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4237875e-d36d-4c45-9102-2f1242407c46-util\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.445497 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" event={"ID":"4237875e-d36d-4c45-9102-2f1242407c46","Type":"ContainerDied","Data":"263a4f6b7c5891ab65cb6a080692b1f34ccb57b6039a8938979e2179caa9a558"} Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.446413 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="263a4f6b7c5891ab65cb6a080692b1f34ccb57b6039a8938979e2179caa9a558" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.445504 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.447913 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-ngsxr" event={"ID":"575383b0-d5a9-41a6-8799-f7cd7e3e637f","Type":"ContainerDied","Data":"3c1d5ae5f9032b239e015d5acf8f0c24300e7577ddc0f168f7c17804a4667fbe"} Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.447948 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c1d5ae5f9032b239e015d5acf8f0c24300e7577ddc0f168f7c17804a4667fbe" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.448056 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-ngsxr" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.632713 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4s625"] Nov 25 22:08:09 crc kubenswrapper[4931]: E1125 22:08:09.633342 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerName="extract" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633364 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerName="extract" Nov 25 22:08:09 crc kubenswrapper[4931]: E1125 22:08:09.633387 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4237875e-d36d-4c45-9102-2f1242407c46" containerName="pull" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633395 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4237875e-d36d-4c45-9102-2f1242407c46" containerName="pull" Nov 25 22:08:09 crc kubenswrapper[4931]: E1125 22:08:09.633406 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4237875e-d36d-4c45-9102-2f1242407c46" containerName="util" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633415 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4237875e-d36d-4c45-9102-2f1242407c46" containerName="util" Nov 25 22:08:09 crc kubenswrapper[4931]: E1125 22:08:09.633428 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="575383b0-d5a9-41a6-8799-f7cd7e3e637f" containerName="keystone-db-sync" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633436 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="575383b0-d5a9-41a6-8799-f7cd7e3e637f" containerName="keystone-db-sync" Nov 25 22:08:09 crc kubenswrapper[4931]: E1125 22:08:09.633449 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerName="pull" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633456 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerName="pull" Nov 25 22:08:09 crc kubenswrapper[4931]: E1125 22:08:09.633465 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4237875e-d36d-4c45-9102-2f1242407c46" containerName="extract" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633473 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4237875e-d36d-4c45-9102-2f1242407c46" containerName="extract" Nov 25 22:08:09 crc kubenswrapper[4931]: E1125 22:08:09.633487 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerName="util" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633495 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerName="util" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633618 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="575383b0-d5a9-41a6-8799-f7cd7e3e637f" containerName="keystone-db-sync" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633631 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a748bed2-f75d-4cfb-b20d-3914cb4094ee" containerName="extract" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.633647 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4237875e-d36d-4c45-9102-2f1242407c46" containerName="extract" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.634134 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.642631 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.642768 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.642936 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.642943 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.643250 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-2j6qp" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.648918 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4s625"] Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.702583 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-credential-keys\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.703039 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-config-data\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.703134 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-fernet-keys\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.703210 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lrj5\" (UniqueName: \"kubernetes.io/projected/200fa711-1e82-4671-8281-555f5139ee24-kube-api-access-6lrj5\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.703301 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-scripts\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.804481 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-scripts\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.804756 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-credential-keys\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.804893 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-config-data\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.804990 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-fernet-keys\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.805116 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lrj5\" (UniqueName: \"kubernetes.io/projected/200fa711-1e82-4671-8281-555f5139ee24-kube-api-access-6lrj5\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.809601 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-credential-keys\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.810009 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-fernet-keys\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.813318 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-config-data\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.815672 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-scripts\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.828513 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lrj5\" (UniqueName: \"kubernetes.io/projected/200fa711-1e82-4671-8281-555f5139ee24-kube-api-access-6lrj5\") pod \"keystone-bootstrap-4s625\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:09 crc kubenswrapper[4931]: I1125 22:08:09.992251 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:10 crc kubenswrapper[4931]: I1125 22:08:10.463984 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4s625"] Nov 25 22:08:11 crc kubenswrapper[4931]: I1125 22:08:11.491142 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4s625" event={"ID":"200fa711-1e82-4671-8281-555f5139ee24","Type":"ContainerStarted","Data":"1685fe75f11d8354658228b93dcf6f13b6b974e4bc838416f7c599769371439e"} Nov 25 22:08:11 crc kubenswrapper[4931]: I1125 22:08:11.491543 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4s625" event={"ID":"200fa711-1e82-4671-8281-555f5139ee24","Type":"ContainerStarted","Data":"6a1c38861c4dab525b6ece650297159c694b3adb97cd5c0692c3f0b3dc95ba99"} Nov 25 22:08:11 crc kubenswrapper[4931]: I1125 22:08:11.520928 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-4s625" podStartSLOduration=2.52090383 podStartE2EDuration="2.52090383s" podCreationTimestamp="2025-11-25 22:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:08:11.513219393 +0000 UTC m=+1029.869731552" watchObservedRunningTime="2025-11-25 22:08:11.52090383 +0000 UTC m=+1029.877415979" Nov 25 22:08:14 crc kubenswrapper[4931]: I1125 22:08:14.512023 4931 generic.go:334] "Generic (PLEG): container finished" podID="200fa711-1e82-4671-8281-555f5139ee24" containerID="1685fe75f11d8354658228b93dcf6f13b6b974e4bc838416f7c599769371439e" exitCode=0 Nov 25 22:08:14 crc kubenswrapper[4931]: I1125 22:08:14.512106 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4s625" event={"ID":"200fa711-1e82-4671-8281-555f5139ee24","Type":"ContainerDied","Data":"1685fe75f11d8354658228b93dcf6f13b6b974e4bc838416f7c599769371439e"} Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.853236 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.907903 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lrj5\" (UniqueName: \"kubernetes.io/projected/200fa711-1e82-4671-8281-555f5139ee24-kube-api-access-6lrj5\") pod \"200fa711-1e82-4671-8281-555f5139ee24\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.907964 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-credential-keys\") pod \"200fa711-1e82-4671-8281-555f5139ee24\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.908003 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-config-data\") pod \"200fa711-1e82-4671-8281-555f5139ee24\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.908044 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-fernet-keys\") pod \"200fa711-1e82-4671-8281-555f5139ee24\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.908074 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-scripts\") pod \"200fa711-1e82-4671-8281-555f5139ee24\" (UID: \"200fa711-1e82-4671-8281-555f5139ee24\") " Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.929468 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-scripts" (OuterVolumeSpecName: "scripts") pod "200fa711-1e82-4671-8281-555f5139ee24" (UID: "200fa711-1e82-4671-8281-555f5139ee24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.930358 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "200fa711-1e82-4671-8281-555f5139ee24" (UID: "200fa711-1e82-4671-8281-555f5139ee24"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.930725 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "200fa711-1e82-4671-8281-555f5139ee24" (UID: "200fa711-1e82-4671-8281-555f5139ee24"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.933308 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/200fa711-1e82-4671-8281-555f5139ee24-kube-api-access-6lrj5" (OuterVolumeSpecName: "kube-api-access-6lrj5") pod "200fa711-1e82-4671-8281-555f5139ee24" (UID: "200fa711-1e82-4671-8281-555f5139ee24"). InnerVolumeSpecName "kube-api-access-6lrj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:08:15 crc kubenswrapper[4931]: I1125 22:08:15.943722 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-config-data" (OuterVolumeSpecName: "config-data") pod "200fa711-1e82-4671-8281-555f5139ee24" (UID: "200fa711-1e82-4671-8281-555f5139ee24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.009234 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.009261 4931 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.009269 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.009278 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lrj5\" (UniqueName: \"kubernetes.io/projected/200fa711-1e82-4671-8281-555f5139ee24-kube-api-access-6lrj5\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.009289 4931 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/200fa711-1e82-4671-8281-555f5139ee24-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.533439 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4s625" event={"ID":"200fa711-1e82-4671-8281-555f5139ee24","Type":"ContainerDied","Data":"6a1c38861c4dab525b6ece650297159c694b3adb97cd5c0692c3f0b3dc95ba99"} Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.533508 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a1c38861c4dab525b6ece650297159c694b3adb97cd5c0692c3f0b3dc95ba99" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.533552 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4s625" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.631861 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7"] Nov 25 22:08:16 crc kubenswrapper[4931]: E1125 22:08:16.632400 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="200fa711-1e82-4671-8281-555f5139ee24" containerName="keystone-bootstrap" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.632483 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="200fa711-1e82-4671-8281-555f5139ee24" containerName="keystone-bootstrap" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.632722 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="200fa711-1e82-4671-8281-555f5139ee24" containerName="keystone-bootstrap" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.633518 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.636323 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.638267 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-2j6qp" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.639402 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.639943 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.650102 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7"] Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.718980 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-credential-keys\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.719536 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-fernet-keys\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.719631 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-scripts\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.719665 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-config-data\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.719772 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmxbb\" (UniqueName: \"kubernetes.io/projected/6e715806-4ba6-436a-b623-1403898fd27a-kube-api-access-fmxbb\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.821551 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-fernet-keys\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.821613 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-scripts\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.821641 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-config-data\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.821675 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmxbb\" (UniqueName: \"kubernetes.io/projected/6e715806-4ba6-436a-b623-1403898fd27a-kube-api-access-fmxbb\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.821707 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-credential-keys\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.827033 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-credential-keys\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.828329 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-scripts\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.832564 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-config-data\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.833454 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e715806-4ba6-436a-b623-1403898fd27a-fernet-keys\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.855647 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmxbb\" (UniqueName: \"kubernetes.io/projected/6e715806-4ba6-436a-b623-1403898fd27a-kube-api-access-fmxbb\") pod \"keystone-8d4cbfbc4-b5kz7\" (UID: \"6e715806-4ba6-436a-b623-1403898fd27a\") " pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:16 crc kubenswrapper[4931]: I1125 22:08:16.954793 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:17 crc kubenswrapper[4931]: I1125 22:08:17.232651 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7"] Nov 25 22:08:17 crc kubenswrapper[4931]: I1125 22:08:17.543208 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" event={"ID":"6e715806-4ba6-436a-b623-1403898fd27a","Type":"ContainerStarted","Data":"98b6e38dd343e633e47e8937e4d15457abeb944dd9cee72dabadbcaea25ce3a0"} Nov 25 22:08:18 crc kubenswrapper[4931]: I1125 22:08:18.557174 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" event={"ID":"6e715806-4ba6-436a-b623-1403898fd27a","Type":"ContainerStarted","Data":"391c39b1e136e74e632d6e7959ca84ed9e46d9519a4158e02a3235e77c211baf"} Nov 25 22:08:18 crc kubenswrapper[4931]: I1125 22:08:18.559280 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:18 crc kubenswrapper[4931]: I1125 22:08:18.584599 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" podStartSLOduration=2.584574259 podStartE2EDuration="2.584574259s" podCreationTimestamp="2025-11-25 22:08:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:08:18.583313755 +0000 UTC m=+1036.939825904" watchObservedRunningTime="2025-11-25 22:08:18.584574259 +0000 UTC m=+1036.941086418" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.404101 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv"] Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.405730 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.410681 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-7wqhd" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.410961 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.430024 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv"] Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.571273 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nppwb\" (UniqueName: \"kubernetes.io/projected/4e0590b1-c897-4382-bc93-4a1b0653a526-kube-api-access-nppwb\") pod \"horizon-operator-controller-manager-6bfcf864bd-cfmpv\" (UID: \"4e0590b1-c897-4382-bc93-4a1b0653a526\") " pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.571586 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e0590b1-c897-4382-bc93-4a1b0653a526-apiservice-cert\") pod \"horizon-operator-controller-manager-6bfcf864bd-cfmpv\" (UID: \"4e0590b1-c897-4382-bc93-4a1b0653a526\") " pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.571613 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e0590b1-c897-4382-bc93-4a1b0653a526-webhook-cert\") pod \"horizon-operator-controller-manager-6bfcf864bd-cfmpv\" (UID: \"4e0590b1-c897-4382-bc93-4a1b0653a526\") " pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.672623 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nppwb\" (UniqueName: \"kubernetes.io/projected/4e0590b1-c897-4382-bc93-4a1b0653a526-kube-api-access-nppwb\") pod \"horizon-operator-controller-manager-6bfcf864bd-cfmpv\" (UID: \"4e0590b1-c897-4382-bc93-4a1b0653a526\") " pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.672672 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e0590b1-c897-4382-bc93-4a1b0653a526-apiservice-cert\") pod \"horizon-operator-controller-manager-6bfcf864bd-cfmpv\" (UID: \"4e0590b1-c897-4382-bc93-4a1b0653a526\") " pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.672702 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e0590b1-c897-4382-bc93-4a1b0653a526-webhook-cert\") pod \"horizon-operator-controller-manager-6bfcf864bd-cfmpv\" (UID: \"4e0590b1-c897-4382-bc93-4a1b0653a526\") " pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.680980 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e0590b1-c897-4382-bc93-4a1b0653a526-webhook-cert\") pod \"horizon-operator-controller-manager-6bfcf864bd-cfmpv\" (UID: \"4e0590b1-c897-4382-bc93-4a1b0653a526\") " pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.681617 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e0590b1-c897-4382-bc93-4a1b0653a526-apiservice-cert\") pod \"horizon-operator-controller-manager-6bfcf864bd-cfmpv\" (UID: \"4e0590b1-c897-4382-bc93-4a1b0653a526\") " pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.693828 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nppwb\" (UniqueName: \"kubernetes.io/projected/4e0590b1-c897-4382-bc93-4a1b0653a526-kube-api-access-nppwb\") pod \"horizon-operator-controller-manager-6bfcf864bd-cfmpv\" (UID: \"4e0590b1-c897-4382-bc93-4a1b0653a526\") " pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:28 crc kubenswrapper[4931]: I1125 22:08:28.727853 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:29 crc kubenswrapper[4931]: I1125 22:08:29.186020 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv"] Nov 25 22:08:29 crc kubenswrapper[4931]: I1125 22:08:29.636262 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" event={"ID":"4e0590b1-c897-4382-bc93-4a1b0653a526","Type":"ContainerStarted","Data":"3b92de4720a9ce87526525c536798373f77fb2a32ad8c4022825e63bd7319257"} Nov 25 22:08:32 crc kubenswrapper[4931]: I1125 22:08:32.656781 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" event={"ID":"4e0590b1-c897-4382-bc93-4a1b0653a526","Type":"ContainerStarted","Data":"af490f90de1d00d98315fa959270aa4c4c0a7da562d4c067edccd71ebd5b709a"} Nov 25 22:08:32 crc kubenswrapper[4931]: I1125 22:08:32.657162 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:32 crc kubenswrapper[4931]: I1125 22:08:32.677630 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" podStartSLOduration=1.740237727 podStartE2EDuration="4.677602602s" podCreationTimestamp="2025-11-25 22:08:28 +0000 UTC" firstStartedPulling="2025-11-25 22:08:29.193452142 +0000 UTC m=+1047.549964251" lastFinishedPulling="2025-11-25 22:08:32.130817007 +0000 UTC m=+1050.487329126" observedRunningTime="2025-11-25 22:08:32.670281105 +0000 UTC m=+1051.026793224" watchObservedRunningTime="2025-11-25 22:08:32.677602602 +0000 UTC m=+1051.034114741" Nov 25 22:08:34 crc kubenswrapper[4931]: I1125 22:08:34.929180 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz"] Nov 25 22:08:34 crc kubenswrapper[4931]: I1125 22:08:34.930269 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:34 crc kubenswrapper[4931]: I1125 22:08:34.933302 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bpgdg" Nov 25 22:08:34 crc kubenswrapper[4931]: I1125 22:08:34.933334 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 25 22:08:34 crc kubenswrapper[4931]: I1125 22:08:34.947377 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz"] Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.103863 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8f67c56d-e216-435c-8384-91d28a4d864e-apiservice-cert\") pod \"swift-operator-controller-manager-7cb66c858-rgrsz\" (UID: \"8f67c56d-e216-435c-8384-91d28a4d864e\") " pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.104233 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8f67c56d-e216-435c-8384-91d28a4d864e-webhook-cert\") pod \"swift-operator-controller-manager-7cb66c858-rgrsz\" (UID: \"8f67c56d-e216-435c-8384-91d28a4d864e\") " pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.104276 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn4fh\" (UniqueName: \"kubernetes.io/projected/8f67c56d-e216-435c-8384-91d28a4d864e-kube-api-access-qn4fh\") pod \"swift-operator-controller-manager-7cb66c858-rgrsz\" (UID: \"8f67c56d-e216-435c-8384-91d28a4d864e\") " pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.205587 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8f67c56d-e216-435c-8384-91d28a4d864e-apiservice-cert\") pod \"swift-operator-controller-manager-7cb66c858-rgrsz\" (UID: \"8f67c56d-e216-435c-8384-91d28a4d864e\") " pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.205630 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8f67c56d-e216-435c-8384-91d28a4d864e-webhook-cert\") pod \"swift-operator-controller-manager-7cb66c858-rgrsz\" (UID: \"8f67c56d-e216-435c-8384-91d28a4d864e\") " pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.205659 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn4fh\" (UniqueName: \"kubernetes.io/projected/8f67c56d-e216-435c-8384-91d28a4d864e-kube-api-access-qn4fh\") pod \"swift-operator-controller-manager-7cb66c858-rgrsz\" (UID: \"8f67c56d-e216-435c-8384-91d28a4d864e\") " pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.212417 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8f67c56d-e216-435c-8384-91d28a4d864e-apiservice-cert\") pod \"swift-operator-controller-manager-7cb66c858-rgrsz\" (UID: \"8f67c56d-e216-435c-8384-91d28a4d864e\") " pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.219233 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8f67c56d-e216-435c-8384-91d28a4d864e-webhook-cert\") pod \"swift-operator-controller-manager-7cb66c858-rgrsz\" (UID: \"8f67c56d-e216-435c-8384-91d28a4d864e\") " pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.223791 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn4fh\" (UniqueName: \"kubernetes.io/projected/8f67c56d-e216-435c-8384-91d28a4d864e-kube-api-access-qn4fh\") pod \"swift-operator-controller-manager-7cb66c858-rgrsz\" (UID: \"8f67c56d-e216-435c-8384-91d28a4d864e\") " pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.252717 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.570772 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz"] Nov 25 22:08:35 crc kubenswrapper[4931]: W1125 22:08:35.580122 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f67c56d_e216_435c_8384_91d28a4d864e.slice/crio-357395c04a81e1ff9db5863b9a05318f23934f9aef97ad37dfe98379464c4953 WatchSource:0}: Error finding container 357395c04a81e1ff9db5863b9a05318f23934f9aef97ad37dfe98379464c4953: Status 404 returned error can't find the container with id 357395c04a81e1ff9db5863b9a05318f23934f9aef97ad37dfe98379464c4953 Nov 25 22:08:35 crc kubenswrapper[4931]: I1125 22:08:35.675488 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" event={"ID":"8f67c56d-e216-435c-8384-91d28a4d864e","Type":"ContainerStarted","Data":"357395c04a81e1ff9db5863b9a05318f23934f9aef97ad37dfe98379464c4953"} Nov 25 22:08:38 crc kubenswrapper[4931]: I1125 22:08:38.699228 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" event={"ID":"8f67c56d-e216-435c-8384-91d28a4d864e","Type":"ContainerStarted","Data":"21a452c30a9e676b35b4edc64252a39b87cab9ce149dfd4d212b6915c586d372"} Nov 25 22:08:38 crc kubenswrapper[4931]: I1125 22:08:38.699643 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:38 crc kubenswrapper[4931]: I1125 22:08:38.716021 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" podStartSLOduration=2.242203807 podStartE2EDuration="4.716004066s" podCreationTimestamp="2025-11-25 22:08:34 +0000 UTC" firstStartedPulling="2025-11-25 22:08:35.581649769 +0000 UTC m=+1053.938161888" lastFinishedPulling="2025-11-25 22:08:38.055450028 +0000 UTC m=+1056.411962147" observedRunningTime="2025-11-25 22:08:38.715518163 +0000 UTC m=+1057.072030282" watchObservedRunningTime="2025-11-25 22:08:38.716004066 +0000 UTC m=+1057.072516185" Nov 25 22:08:38 crc kubenswrapper[4931]: I1125 22:08:38.731794 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:08:38 crc kubenswrapper[4931]: I1125 22:08:38.843308 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:08:38 crc kubenswrapper[4931]: I1125 22:08:38.843364 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:08:38 crc kubenswrapper[4931]: I1125 22:08:38.843404 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 22:08:38 crc kubenswrapper[4931]: I1125 22:08:38.843938 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f284a26c69ccc770f75121f3dbe719003137727be2b936d7ebbd69a2ec135257"} pod="openshift-machine-config-operator/machine-config-daemon-djjhg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 22:08:38 crc kubenswrapper[4931]: I1125 22:08:38.843994 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" containerID="cri-o://f284a26c69ccc770f75121f3dbe719003137727be2b936d7ebbd69a2ec135257" gracePeriod=600 Nov 25 22:08:39 crc kubenswrapper[4931]: I1125 22:08:39.707956 4931 generic.go:334] "Generic (PLEG): container finished" podID="e22448de-5122-43be-853a-081347c13cf2" containerID="f284a26c69ccc770f75121f3dbe719003137727be2b936d7ebbd69a2ec135257" exitCode=0 Nov 25 22:08:39 crc kubenswrapper[4931]: I1125 22:08:39.708000 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerDied","Data":"f284a26c69ccc770f75121f3dbe719003137727be2b936d7ebbd69a2ec135257"} Nov 25 22:08:39 crc kubenswrapper[4931]: I1125 22:08:39.708594 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"9ac1334287ef510421b58306fba1e2a55e64de3a45349db9fac90d16b42c433c"} Nov 25 22:08:39 crc kubenswrapper[4931]: I1125 22:08:39.708637 4931 scope.go:117] "RemoveContainer" containerID="e502eeb6e6462755666e0a9f82f8ea7195e2305ad1bd997e1c4a8dc5d91dd65a" Nov 25 22:08:45 crc kubenswrapper[4931]: I1125 22:08:45.260526 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:08:48 crc kubenswrapper[4931]: I1125 22:08:48.411833 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-8d4cbfbc4-b5kz7" Nov 25 22:08:48 crc kubenswrapper[4931]: I1125 22:08:48.899023 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-svnj9"] Nov 25 22:08:48 crc kubenswrapper[4931]: I1125 22:08:48.899801 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-svnj9" Nov 25 22:08:48 crc kubenswrapper[4931]: I1125 22:08:48.902478 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-6rxnn" Nov 25 22:08:48 crc kubenswrapper[4931]: I1125 22:08:48.915176 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-svnj9"] Nov 25 22:08:48 crc kubenswrapper[4931]: I1125 22:08:48.998171 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rv2w\" (UniqueName: \"kubernetes.io/projected/559c2f5b-32bc-4a0e-95ef-d7f4ad7a902a-kube-api-access-2rv2w\") pod \"glance-operator-index-svnj9\" (UID: \"559c2f5b-32bc-4a0e-95ef-d7f4ad7a902a\") " pod="openstack-operators/glance-operator-index-svnj9" Nov 25 22:08:49 crc kubenswrapper[4931]: I1125 22:08:49.099774 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rv2w\" (UniqueName: \"kubernetes.io/projected/559c2f5b-32bc-4a0e-95ef-d7f4ad7a902a-kube-api-access-2rv2w\") pod \"glance-operator-index-svnj9\" (UID: \"559c2f5b-32bc-4a0e-95ef-d7f4ad7a902a\") " pod="openstack-operators/glance-operator-index-svnj9" Nov 25 22:08:49 crc kubenswrapper[4931]: I1125 22:08:49.132047 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rv2w\" (UniqueName: \"kubernetes.io/projected/559c2f5b-32bc-4a0e-95ef-d7f4ad7a902a-kube-api-access-2rv2w\") pod \"glance-operator-index-svnj9\" (UID: \"559c2f5b-32bc-4a0e-95ef-d7f4ad7a902a\") " pod="openstack-operators/glance-operator-index-svnj9" Nov 25 22:08:49 crc kubenswrapper[4931]: I1125 22:08:49.214210 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-svnj9" Nov 25 22:08:49 crc kubenswrapper[4931]: I1125 22:08:49.703580 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-svnj9"] Nov 25 22:08:49 crc kubenswrapper[4931]: I1125 22:08:49.780306 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-svnj9" event={"ID":"559c2f5b-32bc-4a0e-95ef-d7f4ad7a902a","Type":"ContainerStarted","Data":"c0f581b324530ffc1b5454e8c2d8d5f8dc92c706071b6b16f8f55adf0cf95a69"} Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.057443 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.062807 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.065972 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-gvz2q" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.065978 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.066070 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.076128 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.066307 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.215222 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2f56d793-2122-4466-8dda-38687d951a74-lock\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.215642 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.215665 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2f56d793-2122-4466-8dda-38687d951a74-cache\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.215689 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvx7k\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-kube-api-access-zvx7k\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.215940 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.317197 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2f56d793-2122-4466-8dda-38687d951a74-lock\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.317356 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.317395 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2f56d793-2122-4466-8dda-38687d951a74-cache\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.317451 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvx7k\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-kube-api-access-zvx7k\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: E1125 22:08:50.317514 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:50 crc kubenswrapper[4931]: E1125 22:08:50.317540 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 22:08:50 crc kubenswrapper[4931]: E1125 22:08:50.317600 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift podName:2f56d793-2122-4466-8dda-38687d951a74 nodeName:}" failed. No retries permitted until 2025-11-25 22:08:50.817581295 +0000 UTC m=+1069.174093414 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift") pod "swift-storage-0" (UID: "2f56d793-2122-4466-8dda-38687d951a74") : configmap "swift-ring-files" not found Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.317540 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.317797 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2f56d793-2122-4466-8dda-38687d951a74-lock\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.318050 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.318117 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2f56d793-2122-4466-8dda-38687d951a74-cache\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.337670 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvx7k\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-kube-api-access-zvx7k\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.350305 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.824729 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:50 crc kubenswrapper[4931]: E1125 22:08:50.825000 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:50 crc kubenswrapper[4931]: E1125 22:08:50.825037 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 22:08:50 crc kubenswrapper[4931]: E1125 22:08:50.825098 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift podName:2f56d793-2122-4466-8dda-38687d951a74 nodeName:}" failed. No retries permitted until 2025-11-25 22:08:51.825076793 +0000 UTC m=+1070.181588912 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift") pod "swift-storage-0" (UID: "2f56d793-2122-4466-8dda-38687d951a74") : configmap "swift-ring-files" not found Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.878950 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd"] Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.880174 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.906544 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd"] Nov 25 22:08:50 crc kubenswrapper[4931]: I1125 22:08:50.906781 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.035458 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb416fb6-2f7e-405f-aba4-588fa1929bbd-config-data\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.035496 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb416fb6-2f7e-405f-aba4-588fa1929bbd-run-httpd\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.035538 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrp5v\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-kube-api-access-lrp5v\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.035565 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb416fb6-2f7e-405f-aba4-588fa1929bbd-log-httpd\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.036107 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.137833 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb416fb6-2f7e-405f-aba4-588fa1929bbd-run-httpd\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.138135 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb416fb6-2f7e-405f-aba4-588fa1929bbd-config-data\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.138173 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrp5v\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-kube-api-access-lrp5v\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.138200 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb416fb6-2f7e-405f-aba4-588fa1929bbd-log-httpd\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.138236 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.138369 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb416fb6-2f7e-405f-aba4-588fa1929bbd-run-httpd\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: E1125 22:08:51.138388 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:51 crc kubenswrapper[4931]: E1125 22:08:51.138403 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd: configmap "swift-ring-files" not found Nov 25 22:08:51 crc kubenswrapper[4931]: E1125 22:08:51.138470 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift podName:fb416fb6-2f7e-405f-aba4-588fa1929bbd nodeName:}" failed. No retries permitted until 2025-11-25 22:08:51.638449197 +0000 UTC m=+1069.994961316 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift") pod "swift-proxy-6bd58cfcf7-4j5sd" (UID: "fb416fb6-2f7e-405f-aba4-588fa1929bbd") : configmap "swift-ring-files" not found Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.138648 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb416fb6-2f7e-405f-aba4-588fa1929bbd-log-httpd\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.150286 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb416fb6-2f7e-405f-aba4-588fa1929bbd-config-data\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.154211 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrp5v\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-kube-api-access-lrp5v\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.644729 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:51 crc kubenswrapper[4931]: E1125 22:08:51.644867 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:51 crc kubenswrapper[4931]: E1125 22:08:51.644901 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd: configmap "swift-ring-files" not found Nov 25 22:08:51 crc kubenswrapper[4931]: E1125 22:08:51.644948 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift podName:fb416fb6-2f7e-405f-aba4-588fa1929bbd nodeName:}" failed. No retries permitted until 2025-11-25 22:08:52.644934499 +0000 UTC m=+1071.001446618 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift") pod "swift-proxy-6bd58cfcf7-4j5sd" (UID: "fb416fb6-2f7e-405f-aba4-588fa1929bbd") : configmap "swift-ring-files" not found Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.796123 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-svnj9" event={"ID":"559c2f5b-32bc-4a0e-95ef-d7f4ad7a902a","Type":"ContainerStarted","Data":"5dde0cf5cb59dd86db23681f779ca2c5b555e44936f3a65f389605926180ae66"} Nov 25 22:08:51 crc kubenswrapper[4931]: I1125 22:08:51.846898 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:51 crc kubenswrapper[4931]: E1125 22:08:51.847214 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:51 crc kubenswrapper[4931]: E1125 22:08:51.847233 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 22:08:51 crc kubenswrapper[4931]: E1125 22:08:51.847283 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift podName:2f56d793-2122-4466-8dda-38687d951a74 nodeName:}" failed. No retries permitted until 2025-11-25 22:08:53.847267734 +0000 UTC m=+1072.203779863 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift") pod "swift-storage-0" (UID: "2f56d793-2122-4466-8dda-38687d951a74") : configmap "swift-ring-files" not found Nov 25 22:08:52 crc kubenswrapper[4931]: I1125 22:08:52.657610 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:52 crc kubenswrapper[4931]: E1125 22:08:52.657841 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:52 crc kubenswrapper[4931]: E1125 22:08:52.657900 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd: configmap "swift-ring-files" not found Nov 25 22:08:52 crc kubenswrapper[4931]: E1125 22:08:52.657967 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift podName:fb416fb6-2f7e-405f-aba4-588fa1929bbd nodeName:}" failed. No retries permitted until 2025-11-25 22:08:54.657947812 +0000 UTC m=+1073.014459931 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift") pod "swift-proxy-6bd58cfcf7-4j5sd" (UID: "fb416fb6-2f7e-405f-aba4-588fa1929bbd") : configmap "swift-ring-files" not found Nov 25 22:08:53 crc kubenswrapper[4931]: I1125 22:08:53.874524 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:53 crc kubenswrapper[4931]: E1125 22:08:53.874775 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:53 crc kubenswrapper[4931]: E1125 22:08:53.874818 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 22:08:53 crc kubenswrapper[4931]: E1125 22:08:53.874955 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift podName:2f56d793-2122-4466-8dda-38687d951a74 nodeName:}" failed. No retries permitted until 2025-11-25 22:08:57.874921226 +0000 UTC m=+1076.231433385 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift") pod "swift-storage-0" (UID: "2f56d793-2122-4466-8dda-38687d951a74") : configmap "swift-ring-files" not found Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.025191 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-svnj9" podStartSLOduration=4.283478284 podStartE2EDuration="6.025160379s" podCreationTimestamp="2025-11-25 22:08:48 +0000 UTC" firstStartedPulling="2025-11-25 22:08:49.710341822 +0000 UTC m=+1068.066853941" lastFinishedPulling="2025-11-25 22:08:51.452023907 +0000 UTC m=+1069.808536036" observedRunningTime="2025-11-25 22:08:51.819426165 +0000 UTC m=+1070.175938304" watchObservedRunningTime="2025-11-25 22:08:54.025160379 +0000 UTC m=+1072.381672538" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.026117 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-w2kg9"] Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.027840 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.032331 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.033171 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.039084 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-w2kg9"] Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.179226 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-swiftconf\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.179688 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-dispersionconf\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.179872 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-scripts\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.179977 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/796108c0-b88f-4a48-8596-15155c91b126-etc-swift\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.180047 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-ring-data-devices\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.180112 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4nqs\" (UniqueName: \"kubernetes.io/projected/796108c0-b88f-4a48-8596-15155c91b126-kube-api-access-s4nqs\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.281424 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-scripts\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.281534 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/796108c0-b88f-4a48-8596-15155c91b126-etc-swift\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.281611 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-ring-data-devices\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.282269 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/796108c0-b88f-4a48-8596-15155c91b126-etc-swift\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.282411 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-scripts\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.283460 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-ring-data-devices\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.283634 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4nqs\" (UniqueName: \"kubernetes.io/projected/796108c0-b88f-4a48-8596-15155c91b126-kube-api-access-s4nqs\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.283714 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-swiftconf\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.283746 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-dispersionconf\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.290208 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-swiftconf\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.301563 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-dispersionconf\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.307582 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4nqs\" (UniqueName: \"kubernetes.io/projected/796108c0-b88f-4a48-8596-15155c91b126-kube-api-access-s4nqs\") pod \"swift-ring-rebalance-w2kg9\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.365618 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.692313 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:54 crc kubenswrapper[4931]: E1125 22:08:54.692925 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:54 crc kubenswrapper[4931]: E1125 22:08:54.692966 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd: configmap "swift-ring-files" not found Nov 25 22:08:54 crc kubenswrapper[4931]: E1125 22:08:54.693179 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift podName:fb416fb6-2f7e-405f-aba4-588fa1929bbd nodeName:}" failed. No retries permitted until 2025-11-25 22:08:58.693144217 +0000 UTC m=+1077.049656376 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift") pod "swift-proxy-6bd58cfcf7-4j5sd" (UID: "fb416fb6-2f7e-405f-aba4-588fa1929bbd") : configmap "swift-ring-files" not found Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.795322 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-w2kg9"] Nov 25 22:08:54 crc kubenswrapper[4931]: I1125 22:08:54.818069 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" event={"ID":"796108c0-b88f-4a48-8596-15155c91b126","Type":"ContainerStarted","Data":"6094f634d9a2147c7074d1a59feec9fa02311cc501025e2e0dc00ae4e2081bea"} Nov 25 22:08:57 crc kubenswrapper[4931]: I1125 22:08:57.948847 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:08:57 crc kubenswrapper[4931]: E1125 22:08:57.949056 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:57 crc kubenswrapper[4931]: E1125 22:08:57.949337 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 22:08:57 crc kubenswrapper[4931]: E1125 22:08:57.949396 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift podName:2f56d793-2122-4466-8dda-38687d951a74 nodeName:}" failed. No retries permitted until 2025-11-25 22:09:05.949378324 +0000 UTC m=+1084.305890443 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift") pod "swift-storage-0" (UID: "2f56d793-2122-4466-8dda-38687d951a74") : configmap "swift-ring-files" not found Nov 25 22:08:58 crc kubenswrapper[4931]: I1125 22:08:58.760836 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:08:58 crc kubenswrapper[4931]: E1125 22:08:58.761143 4931 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 22:08:58 crc kubenswrapper[4931]: E1125 22:08:58.761576 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd: configmap "swift-ring-files" not found Nov 25 22:08:58 crc kubenswrapper[4931]: E1125 22:08:58.761712 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift podName:fb416fb6-2f7e-405f-aba4-588fa1929bbd nodeName:}" failed. No retries permitted until 2025-11-25 22:09:06.761692656 +0000 UTC m=+1085.118204835 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift") pod "swift-proxy-6bd58cfcf7-4j5sd" (UID: "fb416fb6-2f7e-405f-aba4-588fa1929bbd") : configmap "swift-ring-files" not found Nov 25 22:08:58 crc kubenswrapper[4931]: I1125 22:08:58.839758 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" event={"ID":"796108c0-b88f-4a48-8596-15155c91b126","Type":"ContainerStarted","Data":"f0eb811cbeb13e7e9cbc4d17f1f8580adddb07462569b17926aec37f3e4b3002"} Nov 25 22:08:58 crc kubenswrapper[4931]: I1125 22:08:58.866409 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" podStartSLOduration=1.204423578 podStartE2EDuration="4.866393344s" podCreationTimestamp="2025-11-25 22:08:54 +0000 UTC" firstStartedPulling="2025-11-25 22:08:54.800329722 +0000 UTC m=+1073.156841881" lastFinishedPulling="2025-11-25 22:08:58.462299518 +0000 UTC m=+1076.818811647" observedRunningTime="2025-11-25 22:08:58.866124326 +0000 UTC m=+1077.222636445" watchObservedRunningTime="2025-11-25 22:08:58.866393344 +0000 UTC m=+1077.222905463" Nov 25 22:08:59 crc kubenswrapper[4931]: I1125 22:08:59.215697 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-svnj9" Nov 25 22:08:59 crc kubenswrapper[4931]: I1125 22:08:59.215744 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-svnj9" Nov 25 22:08:59 crc kubenswrapper[4931]: I1125 22:08:59.248761 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-svnj9" Nov 25 22:08:59 crc kubenswrapper[4931]: I1125 22:08:59.888203 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-svnj9" Nov 25 22:09:03 crc kubenswrapper[4931]: I1125 22:09:03.952867 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq"] Nov 25 22:09:03 crc kubenswrapper[4931]: I1125 22:09:03.955590 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:03 crc kubenswrapper[4931]: I1125 22:09:03.957980 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wrsb7" Nov 25 22:09:03 crc kubenswrapper[4931]: I1125 22:09:03.961242 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq"] Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.044760 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-bundle\") pod \"5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.044870 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-util\") pod \"5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.044950 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29xz7\" (UniqueName: \"kubernetes.io/projected/bdfe8e93-8490-4958-aab7-0c2b062b22fa-kube-api-access-29xz7\") pod \"5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.146543 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29xz7\" (UniqueName: \"kubernetes.io/projected/bdfe8e93-8490-4958-aab7-0c2b062b22fa-kube-api-access-29xz7\") pod \"5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.146720 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-bundle\") pod \"5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.146788 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-util\") pod \"5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.147443 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-util\") pod \"5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.147479 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-bundle\") pod \"5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.189478 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29xz7\" (UniqueName: \"kubernetes.io/projected/bdfe8e93-8490-4958-aab7-0c2b062b22fa-kube-api-access-29xz7\") pod \"5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.276032 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.749954 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq"] Nov 25 22:09:04 crc kubenswrapper[4931]: W1125 22:09:04.761019 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdfe8e93_8490_4958_aab7_0c2b062b22fa.slice/crio-e1cd10ec84dfe450e1130e3514f5fb3f868bf58340ef0e3402f2b8237c24e835 WatchSource:0}: Error finding container e1cd10ec84dfe450e1130e3514f5fb3f868bf58340ef0e3402f2b8237c24e835: Status 404 returned error can't find the container with id e1cd10ec84dfe450e1130e3514f5fb3f868bf58340ef0e3402f2b8237c24e835 Nov 25 22:09:04 crc kubenswrapper[4931]: I1125 22:09:04.902586 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" event={"ID":"bdfe8e93-8490-4958-aab7-0c2b062b22fa","Type":"ContainerStarted","Data":"e1cd10ec84dfe450e1130e3514f5fb3f868bf58340ef0e3402f2b8237c24e835"} Nov 25 22:09:05 crc kubenswrapper[4931]: I1125 22:09:05.910202 4931 generic.go:334] "Generic (PLEG): container finished" podID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerID="b6e73944251fa26af3f789d9cde4d5e8364476f66bc125213c92a57cf44e980c" exitCode=0 Nov 25 22:09:05 crc kubenswrapper[4931]: I1125 22:09:05.910261 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" event={"ID":"bdfe8e93-8490-4958-aab7-0c2b062b22fa","Type":"ContainerDied","Data":"b6e73944251fa26af3f789d9cde4d5e8364476f66bc125213c92a57cf44e980c"} Nov 25 22:09:05 crc kubenswrapper[4931]: I1125 22:09:05.912124 4931 generic.go:334] "Generic (PLEG): container finished" podID="796108c0-b88f-4a48-8596-15155c91b126" containerID="f0eb811cbeb13e7e9cbc4d17f1f8580adddb07462569b17926aec37f3e4b3002" exitCode=0 Nov 25 22:09:05 crc kubenswrapper[4931]: I1125 22:09:05.912158 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" event={"ID":"796108c0-b88f-4a48-8596-15155c91b126","Type":"ContainerDied","Data":"f0eb811cbeb13e7e9cbc4d17f1f8580adddb07462569b17926aec37f3e4b3002"} Nov 25 22:09:05 crc kubenswrapper[4931]: I1125 22:09:05.974842 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:09:05 crc kubenswrapper[4931]: I1125 22:09:05.984244 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2f56d793-2122-4466-8dda-38687d951a74-etc-swift\") pod \"swift-storage-0\" (UID: \"2f56d793-2122-4466-8dda-38687d951a74\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:09:05 crc kubenswrapper[4931]: I1125 22:09:05.999162 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 22:09:06 crc kubenswrapper[4931]: I1125 22:09:06.438749 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 22:09:06 crc kubenswrapper[4931]: W1125 22:09:06.454754 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f56d793_2122_4466_8dda_38687d951a74.slice/crio-c911fbb73c3a8b9e8b908167ebbe72c77fb5a88ea469f3b31c1d18b0e32a576e WatchSource:0}: Error finding container c911fbb73c3a8b9e8b908167ebbe72c77fb5a88ea469f3b31c1d18b0e32a576e: Status 404 returned error can't find the container with id c911fbb73c3a8b9e8b908167ebbe72c77fb5a88ea469f3b31c1d18b0e32a576e Nov 25 22:09:06 crc kubenswrapper[4931]: I1125 22:09:06.789235 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:09:06 crc kubenswrapper[4931]: I1125 22:09:06.798826 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb416fb6-2f7e-405f-aba4-588fa1929bbd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4j5sd\" (UID: \"fb416fb6-2f7e-405f-aba4-588fa1929bbd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:09:06 crc kubenswrapper[4931]: I1125 22:09:06.835501 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:09:06 crc kubenswrapper[4931]: I1125 22:09:06.925272 4931 generic.go:334] "Generic (PLEG): container finished" podID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerID="d7ea4cb4051573058f974cb933bbb64346451b30e14f816eeb14b011392529a5" exitCode=0 Nov 25 22:09:06 crc kubenswrapper[4931]: I1125 22:09:06.925376 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" event={"ID":"bdfe8e93-8490-4958-aab7-0c2b062b22fa","Type":"ContainerDied","Data":"d7ea4cb4051573058f974cb933bbb64346451b30e14f816eeb14b011392529a5"} Nov 25 22:09:06 crc kubenswrapper[4931]: I1125 22:09:06.928280 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"c911fbb73c3a8b9e8b908167ebbe72c77fb5a88ea469f3b31c1d18b0e32a576e"} Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.321019 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd"] Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.340851 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.504224 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-scripts\") pod \"796108c0-b88f-4a48-8596-15155c91b126\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.504320 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-dispersionconf\") pod \"796108c0-b88f-4a48-8596-15155c91b126\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.504352 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4nqs\" (UniqueName: \"kubernetes.io/projected/796108c0-b88f-4a48-8596-15155c91b126-kube-api-access-s4nqs\") pod \"796108c0-b88f-4a48-8596-15155c91b126\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.504389 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-ring-data-devices\") pod \"796108c0-b88f-4a48-8596-15155c91b126\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.504417 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-swiftconf\") pod \"796108c0-b88f-4a48-8596-15155c91b126\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.504443 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/796108c0-b88f-4a48-8596-15155c91b126-etc-swift\") pod \"796108c0-b88f-4a48-8596-15155c91b126\" (UID: \"796108c0-b88f-4a48-8596-15155c91b126\") " Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.505160 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "796108c0-b88f-4a48-8596-15155c91b126" (UID: "796108c0-b88f-4a48-8596-15155c91b126"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.506219 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/796108c0-b88f-4a48-8596-15155c91b126-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "796108c0-b88f-4a48-8596-15155c91b126" (UID: "796108c0-b88f-4a48-8596-15155c91b126"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.509454 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/796108c0-b88f-4a48-8596-15155c91b126-kube-api-access-s4nqs" (OuterVolumeSpecName: "kube-api-access-s4nqs") pod "796108c0-b88f-4a48-8596-15155c91b126" (UID: "796108c0-b88f-4a48-8596-15155c91b126"). InnerVolumeSpecName "kube-api-access-s4nqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.524476 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "796108c0-b88f-4a48-8596-15155c91b126" (UID: "796108c0-b88f-4a48-8596-15155c91b126"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.534692 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "796108c0-b88f-4a48-8596-15155c91b126" (UID: "796108c0-b88f-4a48-8596-15155c91b126"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.539708 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-scripts" (OuterVolumeSpecName: "scripts") pod "796108c0-b88f-4a48-8596-15155c91b126" (UID: "796108c0-b88f-4a48-8596-15155c91b126"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:09:07 crc kubenswrapper[4931]: W1125 22:09:07.588413 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb416fb6_2f7e_405f_aba4_588fa1929bbd.slice/crio-b188462bd214ac97912fff3ac4e4d9e049fac4bf3269ed0917d0b806aaeace02 WatchSource:0}: Error finding container b188462bd214ac97912fff3ac4e4d9e049fac4bf3269ed0917d0b806aaeace02: Status 404 returned error can't find the container with id b188462bd214ac97912fff3ac4e4d9e049fac4bf3269ed0917d0b806aaeace02 Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.605780 4931 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.605809 4931 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/796108c0-b88f-4a48-8596-15155c91b126-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.605819 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.605827 4931 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/796108c0-b88f-4a48-8596-15155c91b126-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.605837 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4nqs\" (UniqueName: \"kubernetes.io/projected/796108c0-b88f-4a48-8596-15155c91b126-kube-api-access-s4nqs\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.605846 4931 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/796108c0-b88f-4a48-8596-15155c91b126-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.969833 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" event={"ID":"fb416fb6-2f7e-405f-aba4-588fa1929bbd","Type":"ContainerStarted","Data":"509e5d301beb2db255d7e4821640b8a10064b9578284cd389172b2a9ef79bcf8"} Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.969903 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" event={"ID":"fb416fb6-2f7e-405f-aba4-588fa1929bbd","Type":"ContainerStarted","Data":"b188462bd214ac97912fff3ac4e4d9e049fac4bf3269ed0917d0b806aaeace02"} Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.977590 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.977589 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-w2kg9" event={"ID":"796108c0-b88f-4a48-8596-15155c91b126","Type":"ContainerDied","Data":"6094f634d9a2147c7074d1a59feec9fa02311cc501025e2e0dc00ae4e2081bea"} Nov 25 22:09:07 crc kubenswrapper[4931]: I1125 22:09:07.977717 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6094f634d9a2147c7074d1a59feec9fa02311cc501025e2e0dc00ae4e2081bea" Nov 25 22:09:08 crc kubenswrapper[4931]: I1125 22:09:08.004719 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"d265a20c58d6f98f9f2d7c6ee05cac385ed1181873a4c77b0c0a7fee72624b23"} Nov 25 22:09:08 crc kubenswrapper[4931]: I1125 22:09:08.012051 4931 generic.go:334] "Generic (PLEG): container finished" podID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerID="6f356eaed22facc11feb19d63e677b69953de8d944d6979e3439445598bfadf0" exitCode=0 Nov 25 22:09:08 crc kubenswrapper[4931]: I1125 22:09:08.012091 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" event={"ID":"bdfe8e93-8490-4958-aab7-0c2b062b22fa","Type":"ContainerDied","Data":"6f356eaed22facc11feb19d63e677b69953de8d944d6979e3439445598bfadf0"} Nov 25 22:09:08 crc kubenswrapper[4931]: E1125 22:09:08.068815 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod796108c0_b88f_4a48_8596_15155c91b126.slice/crio-6094f634d9a2147c7074d1a59feec9fa02311cc501025e2e0dc00ae4e2081bea\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod796108c0_b88f_4a48_8596_15155c91b126.slice\": RecentStats: unable to find data in memory cache]" Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.020243 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" event={"ID":"fb416fb6-2f7e-405f-aba4-588fa1929bbd","Type":"ContainerStarted","Data":"338cdfaf5e6e880614e0af302e1d31e43fe853c593bacd7cd5b1d8aed02a3534"} Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.020514 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.022949 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"70bd4a7e7d840af8b4512e6ac1530ce053825c00f9468432db8f7d549ab48906"} Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.022985 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"d4ba1db8584c47c547f6f9ed3a14eddea7ea18e5c78aa557862bd9090cde2e0c"} Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.022997 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"bcd3a6451e62117e788b24e61540a9fcbdd707ab72b2bb174947082d9bab5411"} Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.039191 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" podStartSLOduration=19.039175799 podStartE2EDuration="19.039175799s" podCreationTimestamp="2025-11-25 22:08:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:09:09.035176752 +0000 UTC m=+1087.391688871" watchObservedRunningTime="2025-11-25 22:09:09.039175799 +0000 UTC m=+1087.395687918" Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.449287 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.532065 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-util\") pod \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.532132 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-bundle\") pod \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.532217 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29xz7\" (UniqueName: \"kubernetes.io/projected/bdfe8e93-8490-4958-aab7-0c2b062b22fa-kube-api-access-29xz7\") pod \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\" (UID: \"bdfe8e93-8490-4958-aab7-0c2b062b22fa\") " Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.532946 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-bundle" (OuterVolumeSpecName: "bundle") pod "bdfe8e93-8490-4958-aab7-0c2b062b22fa" (UID: "bdfe8e93-8490-4958-aab7-0c2b062b22fa"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.538899 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdfe8e93-8490-4958-aab7-0c2b062b22fa-kube-api-access-29xz7" (OuterVolumeSpecName: "kube-api-access-29xz7") pod "bdfe8e93-8490-4958-aab7-0c2b062b22fa" (UID: "bdfe8e93-8490-4958-aab7-0c2b062b22fa"). InnerVolumeSpecName "kube-api-access-29xz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.548338 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-util" (OuterVolumeSpecName: "util") pod "bdfe8e93-8490-4958-aab7-0c2b062b22fa" (UID: "bdfe8e93-8490-4958-aab7-0c2b062b22fa"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.634379 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-util\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.634429 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bdfe8e93-8490-4958-aab7-0c2b062b22fa-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:09 crc kubenswrapper[4931]: I1125 22:09:09.634450 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29xz7\" (UniqueName: \"kubernetes.io/projected/bdfe8e93-8490-4958-aab7-0c2b062b22fa-kube-api-access-29xz7\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:10 crc kubenswrapper[4931]: I1125 22:09:10.038999 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" event={"ID":"bdfe8e93-8490-4958-aab7-0c2b062b22fa","Type":"ContainerDied","Data":"e1cd10ec84dfe450e1130e3514f5fb3f868bf58340ef0e3402f2b8237c24e835"} Nov 25 22:09:10 crc kubenswrapper[4931]: I1125 22:09:10.039365 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1cd10ec84dfe450e1130e3514f5fb3f868bf58340ef0e3402f2b8237c24e835" Nov 25 22:09:10 crc kubenswrapper[4931]: I1125 22:09:10.039043 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq" Nov 25 22:09:10 crc kubenswrapper[4931]: I1125 22:09:10.046066 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"780cbc05f7f1603d30b91ca234a07cd82947abda9e07c5141d9614f1958d88e3"} Nov 25 22:09:10 crc kubenswrapper[4931]: I1125 22:09:10.046108 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"55b8c3902ac281dbe729db6a2578135b7e8b6f16a3f93af96b9db8258f64bc83"} Nov 25 22:09:10 crc kubenswrapper[4931]: I1125 22:09:10.046209 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:09:10 crc kubenswrapper[4931]: I1125 22:09:10.046235 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"32e5eae3bbff14bb0865e9a4bb8865de2c115ace9cc55b1cb9a81660583285dc"} Nov 25 22:09:10 crc kubenswrapper[4931]: I1125 22:09:10.046250 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"fb2ac63d8f6ebcf9b55e92f54874ee1c5c887fdada32af77e811efb1d375de4c"} Nov 25 22:09:12 crc kubenswrapper[4931]: I1125 22:09:12.078019 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"5c19b4d7a45cdaeb89b7c0a5a7ec1d6a4e1b747d833a917d792bb3859e1f06cc"} Nov 25 22:09:12 crc kubenswrapper[4931]: I1125 22:09:12.078638 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"a123257a8b6000b0ad64689a14f10b59bcc42293b4799661fe159ada3d57d3ed"} Nov 25 22:09:12 crc kubenswrapper[4931]: I1125 22:09:12.078660 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"59baebe4a656cf94603450cbd2eb1663c532ec43cb6e7429367147f56abc3b3b"} Nov 25 22:09:13 crc kubenswrapper[4931]: I1125 22:09:13.096460 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"c79e916808c40acc7d8bbfd41a9f7b4b88007a8d83b4f0cf3c8be551da1cda33"} Nov 25 22:09:13 crc kubenswrapper[4931]: I1125 22:09:13.098152 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"52f8c7c05afa92ef50d5f3099acb65551c482fa6e28667d793e85d24cf4b86ab"} Nov 25 22:09:13 crc kubenswrapper[4931]: I1125 22:09:13.098267 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"68044ea92a731d010fb81973ff4b14d6f24e54105ce45d10f997b85c99cf8152"} Nov 25 22:09:13 crc kubenswrapper[4931]: I1125 22:09:13.098356 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"2f56d793-2122-4466-8dda-38687d951a74","Type":"ContainerStarted","Data":"fafe698aa16da70e87c89bf4f4cfd139cd00c244fbda9964077336332f0b226c"} Nov 25 22:09:13 crc kubenswrapper[4931]: I1125 22:09:13.139803 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=19.236927419 podStartE2EDuration="24.139772072s" podCreationTimestamp="2025-11-25 22:08:49 +0000 UTC" firstStartedPulling="2025-11-25 22:09:06.45759849 +0000 UTC m=+1084.814110609" lastFinishedPulling="2025-11-25 22:09:11.360443123 +0000 UTC m=+1089.716955262" observedRunningTime="2025-11-25 22:09:13.131201261 +0000 UTC m=+1091.487713420" watchObservedRunningTime="2025-11-25 22:09:13.139772072 +0000 UTC m=+1091.496284191" Nov 25 22:09:16 crc kubenswrapper[4931]: I1125 22:09:16.839020 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:09:16 crc kubenswrapper[4931]: I1125 22:09:16.845557 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4j5sd" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.155465 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg"] Nov 25 22:09:20 crc kubenswrapper[4931]: E1125 22:09:20.155976 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796108c0-b88f-4a48-8596-15155c91b126" containerName="swift-ring-rebalance" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.155987 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="796108c0-b88f-4a48-8596-15155c91b126" containerName="swift-ring-rebalance" Nov 25 22:09:20 crc kubenswrapper[4931]: E1125 22:09:20.155997 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerName="util" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.156003 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerName="util" Nov 25 22:09:20 crc kubenswrapper[4931]: E1125 22:09:20.156020 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerName="extract" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.156026 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerName="extract" Nov 25 22:09:20 crc kubenswrapper[4931]: E1125 22:09:20.156035 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerName="pull" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.156041 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerName="pull" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.156171 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="796108c0-b88f-4a48-8596-15155c91b126" containerName="swift-ring-rebalance" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.156181 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdfe8e93-8490-4958-aab7-0c2b062b22fa" containerName="extract" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.156608 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.158531 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.160108 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-mw2wd" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.177245 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg"] Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.302635 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttwvr\" (UniqueName: \"kubernetes.io/projected/0700c2d1-0a31-4d46-9fe7-4bdb809d1cef-kube-api-access-ttwvr\") pod \"glance-operator-controller-manager-6fb5bb54f-chfmg\" (UID: \"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef\") " pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.302683 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0700c2d1-0a31-4d46-9fe7-4bdb809d1cef-webhook-cert\") pod \"glance-operator-controller-manager-6fb5bb54f-chfmg\" (UID: \"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef\") " pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.303022 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0700c2d1-0a31-4d46-9fe7-4bdb809d1cef-apiservice-cert\") pod \"glance-operator-controller-manager-6fb5bb54f-chfmg\" (UID: \"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef\") " pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.404272 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0700c2d1-0a31-4d46-9fe7-4bdb809d1cef-apiservice-cert\") pod \"glance-operator-controller-manager-6fb5bb54f-chfmg\" (UID: \"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef\") " pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.404364 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttwvr\" (UniqueName: \"kubernetes.io/projected/0700c2d1-0a31-4d46-9fe7-4bdb809d1cef-kube-api-access-ttwvr\") pod \"glance-operator-controller-manager-6fb5bb54f-chfmg\" (UID: \"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef\") " pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.404398 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0700c2d1-0a31-4d46-9fe7-4bdb809d1cef-webhook-cert\") pod \"glance-operator-controller-manager-6fb5bb54f-chfmg\" (UID: \"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef\") " pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.413993 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0700c2d1-0a31-4d46-9fe7-4bdb809d1cef-webhook-cert\") pod \"glance-operator-controller-manager-6fb5bb54f-chfmg\" (UID: \"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef\") " pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.414008 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0700c2d1-0a31-4d46-9fe7-4bdb809d1cef-apiservice-cert\") pod \"glance-operator-controller-manager-6fb5bb54f-chfmg\" (UID: \"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef\") " pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.426105 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttwvr\" (UniqueName: \"kubernetes.io/projected/0700c2d1-0a31-4d46-9fe7-4bdb809d1cef-kube-api-access-ttwvr\") pod \"glance-operator-controller-manager-6fb5bb54f-chfmg\" (UID: \"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef\") " pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.474969 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:20 crc kubenswrapper[4931]: I1125 22:09:20.922813 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg"] Nov 25 22:09:20 crc kubenswrapper[4931]: W1125 22:09:20.935977 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0700c2d1_0a31_4d46_9fe7_4bdb809d1cef.slice/crio-32aef84fc6238b85d5bc416d8fe778e30b80eeb52ff488ae539ec4a6d2cbc12c WatchSource:0}: Error finding container 32aef84fc6238b85d5bc416d8fe778e30b80eeb52ff488ae539ec4a6d2cbc12c: Status 404 returned error can't find the container with id 32aef84fc6238b85d5bc416d8fe778e30b80eeb52ff488ae539ec4a6d2cbc12c Nov 25 22:09:21 crc kubenswrapper[4931]: I1125 22:09:21.174917 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" event={"ID":"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef","Type":"ContainerStarted","Data":"32aef84fc6238b85d5bc416d8fe778e30b80eeb52ff488ae539ec4a6d2cbc12c"} Nov 25 22:09:23 crc kubenswrapper[4931]: I1125 22:09:23.188157 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" event={"ID":"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef","Type":"ContainerStarted","Data":"4908a63e7b813b51872992e0b6f626b33e2adac29589994682430011e21c8be5"} Nov 25 22:09:23 crc kubenswrapper[4931]: I1125 22:09:23.188443 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:23 crc kubenswrapper[4931]: I1125 22:09:23.203172 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" podStartSLOduration=1.245358451 podStartE2EDuration="3.203155312s" podCreationTimestamp="2025-11-25 22:09:20 +0000 UTC" firstStartedPulling="2025-11-25 22:09:20.939618613 +0000 UTC m=+1099.296130732" lastFinishedPulling="2025-11-25 22:09:22.897415474 +0000 UTC m=+1101.253927593" observedRunningTime="2025-11-25 22:09:23.201424746 +0000 UTC m=+1101.557936865" watchObservedRunningTime="2025-11-25 22:09:23.203155312 +0000 UTC m=+1101.559667431" Nov 25 22:09:30 crc kubenswrapper[4931]: I1125 22:09:30.481732 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.654955 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-311e-account-create-update-f5k5q"] Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.657018 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.660339 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.661551 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-x8gms"] Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.662915 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.668848 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.669991 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.671726 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-x8gms"] Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.675050 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.675153 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.675251 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.675504 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-zll5q" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.677371 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-311e-account-create-update-f5k5q"] Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.683625 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.779141 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e5f137dc-0889-45c9-bfb7-f99227ae93df-openstack-config\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.779613 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e5f137dc-0889-45c9-bfb7-f99227ae93df-openstack-config-secret\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.779762 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/e5f137dc-0889-45c9-bfb7-f99227ae93df-openstack-scripts\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.779854 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84edc32e-99b0-4d11-8aff-b445d6a03786-operator-scripts\") pod \"glance-db-create-x8gms\" (UID: \"84edc32e-99b0-4d11-8aff-b445d6a03786\") " pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.779995 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6228ec7-9ffd-446c-b51c-99414b57969d-operator-scripts\") pod \"glance-311e-account-create-update-f5k5q\" (UID: \"d6228ec7-9ffd-446c-b51c-99414b57969d\") " pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.780122 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4pbx\" (UniqueName: \"kubernetes.io/projected/d6228ec7-9ffd-446c-b51c-99414b57969d-kube-api-access-z4pbx\") pod \"glance-311e-account-create-update-f5k5q\" (UID: \"d6228ec7-9ffd-446c-b51c-99414b57969d\") " pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.780216 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz7p5\" (UniqueName: \"kubernetes.io/projected/84edc32e-99b0-4d11-8aff-b445d6a03786-kube-api-access-zz7p5\") pod \"glance-db-create-x8gms\" (UID: \"84edc32e-99b0-4d11-8aff-b445d6a03786\") " pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.780300 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppvjs\" (UniqueName: \"kubernetes.io/projected/e5f137dc-0889-45c9-bfb7-f99227ae93df-kube-api-access-ppvjs\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.882239 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6228ec7-9ffd-446c-b51c-99414b57969d-operator-scripts\") pod \"glance-311e-account-create-update-f5k5q\" (UID: \"d6228ec7-9ffd-446c-b51c-99414b57969d\") " pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.882322 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4pbx\" (UniqueName: \"kubernetes.io/projected/d6228ec7-9ffd-446c-b51c-99414b57969d-kube-api-access-z4pbx\") pod \"glance-311e-account-create-update-f5k5q\" (UID: \"d6228ec7-9ffd-446c-b51c-99414b57969d\") " pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.882370 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz7p5\" (UniqueName: \"kubernetes.io/projected/84edc32e-99b0-4d11-8aff-b445d6a03786-kube-api-access-zz7p5\") pod \"glance-db-create-x8gms\" (UID: \"84edc32e-99b0-4d11-8aff-b445d6a03786\") " pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.882407 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppvjs\" (UniqueName: \"kubernetes.io/projected/e5f137dc-0889-45c9-bfb7-f99227ae93df-kube-api-access-ppvjs\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.882510 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e5f137dc-0889-45c9-bfb7-f99227ae93df-openstack-config\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.882562 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e5f137dc-0889-45c9-bfb7-f99227ae93df-openstack-config-secret\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.882630 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/e5f137dc-0889-45c9-bfb7-f99227ae93df-openstack-scripts\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.882672 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84edc32e-99b0-4d11-8aff-b445d6a03786-operator-scripts\") pod \"glance-db-create-x8gms\" (UID: \"84edc32e-99b0-4d11-8aff-b445d6a03786\") " pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.884518 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6228ec7-9ffd-446c-b51c-99414b57969d-operator-scripts\") pod \"glance-311e-account-create-update-f5k5q\" (UID: \"d6228ec7-9ffd-446c-b51c-99414b57969d\") " pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.884575 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e5f137dc-0889-45c9-bfb7-f99227ae93df-openstack-config\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.884664 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/e5f137dc-0889-45c9-bfb7-f99227ae93df-openstack-scripts\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.885097 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84edc32e-99b0-4d11-8aff-b445d6a03786-operator-scripts\") pod \"glance-db-create-x8gms\" (UID: \"84edc32e-99b0-4d11-8aff-b445d6a03786\") " pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.901781 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e5f137dc-0889-45c9-bfb7-f99227ae93df-openstack-config-secret\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.907439 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppvjs\" (UniqueName: \"kubernetes.io/projected/e5f137dc-0889-45c9-bfb7-f99227ae93df-kube-api-access-ppvjs\") pod \"openstackclient\" (UID: \"e5f137dc-0889-45c9-bfb7-f99227ae93df\") " pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.911386 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4pbx\" (UniqueName: \"kubernetes.io/projected/d6228ec7-9ffd-446c-b51c-99414b57969d-kube-api-access-z4pbx\") pod \"glance-311e-account-create-update-f5k5q\" (UID: \"d6228ec7-9ffd-446c-b51c-99414b57969d\") " pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.912197 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz7p5\" (UniqueName: \"kubernetes.io/projected/84edc32e-99b0-4d11-8aff-b445d6a03786-kube-api-access-zz7p5\") pod \"glance-db-create-x8gms\" (UID: \"84edc32e-99b0-4d11-8aff-b445d6a03786\") " pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:36 crc kubenswrapper[4931]: I1125 22:09:36.979298 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:37 crc kubenswrapper[4931]: I1125 22:09:37.000353 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:37 crc kubenswrapper[4931]: I1125 22:09:37.011171 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 22:09:37 crc kubenswrapper[4931]: I1125 22:09:37.492189 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-311e-account-create-update-f5k5q"] Nov 25 22:09:37 crc kubenswrapper[4931]: W1125 22:09:37.495496 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6228ec7_9ffd_446c_b51c_99414b57969d.slice/crio-438d4313f87b37c7d1640002c8f08b8c4ce863b2db54abdd15d9bd92bb37777f WatchSource:0}: Error finding container 438d4313f87b37c7d1640002c8f08b8c4ce863b2db54abdd15d9bd92bb37777f: Status 404 returned error can't find the container with id 438d4313f87b37c7d1640002c8f08b8c4ce863b2db54abdd15d9bd92bb37777f Nov 25 22:09:37 crc kubenswrapper[4931]: I1125 22:09:37.548586 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-x8gms"] Nov 25 22:09:37 crc kubenswrapper[4931]: I1125 22:09:37.554445 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 22:09:37 crc kubenswrapper[4931]: W1125 22:09:37.558160 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84edc32e_99b0_4d11_8aff_b445d6a03786.slice/crio-18d116ac29f7c6e3ca3c819fdd76a47bdeee66f080017dce21e5174de0e6ff4b WatchSource:0}: Error finding container 18d116ac29f7c6e3ca3c819fdd76a47bdeee66f080017dce21e5174de0e6ff4b: Status 404 returned error can't find the container with id 18d116ac29f7c6e3ca3c819fdd76a47bdeee66f080017dce21e5174de0e6ff4b Nov 25 22:09:38 crc kubenswrapper[4931]: I1125 22:09:38.322955 4931 generic.go:334] "Generic (PLEG): container finished" podID="d6228ec7-9ffd-446c-b51c-99414b57969d" containerID="dc58e829c164e2f7e839ab52fafa5e3fca2e525d7a2172092ca55b8f7d30a019" exitCode=0 Nov 25 22:09:38 crc kubenswrapper[4931]: I1125 22:09:38.323054 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" event={"ID":"d6228ec7-9ffd-446c-b51c-99414b57969d","Type":"ContainerDied","Data":"dc58e829c164e2f7e839ab52fafa5e3fca2e525d7a2172092ca55b8f7d30a019"} Nov 25 22:09:38 crc kubenswrapper[4931]: I1125 22:09:38.323377 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" event={"ID":"d6228ec7-9ffd-446c-b51c-99414b57969d","Type":"ContainerStarted","Data":"438d4313f87b37c7d1640002c8f08b8c4ce863b2db54abdd15d9bd92bb37777f"} Nov 25 22:09:38 crc kubenswrapper[4931]: I1125 22:09:38.327022 4931 generic.go:334] "Generic (PLEG): container finished" podID="84edc32e-99b0-4d11-8aff-b445d6a03786" containerID="3ca62f88382d274d24949e6f4a8a15e886ad177d4c690a3ac937f185773144e8" exitCode=0 Nov 25 22:09:38 crc kubenswrapper[4931]: I1125 22:09:38.327111 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-x8gms" event={"ID":"84edc32e-99b0-4d11-8aff-b445d6a03786","Type":"ContainerDied","Data":"3ca62f88382d274d24949e6f4a8a15e886ad177d4c690a3ac937f185773144e8"} Nov 25 22:09:38 crc kubenswrapper[4931]: I1125 22:09:38.327208 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-x8gms" event={"ID":"84edc32e-99b0-4d11-8aff-b445d6a03786","Type":"ContainerStarted","Data":"18d116ac29f7c6e3ca3c819fdd76a47bdeee66f080017dce21e5174de0e6ff4b"} Nov 25 22:09:38 crc kubenswrapper[4931]: I1125 22:09:38.329172 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"e5f137dc-0889-45c9-bfb7-f99227ae93df","Type":"ContainerStarted","Data":"2a50a754a9a962746b16ca441eedcc6f9653934cede405189fdcaf219b895ad1"} Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.682210 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.688736 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.829190 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4pbx\" (UniqueName: \"kubernetes.io/projected/d6228ec7-9ffd-446c-b51c-99414b57969d-kube-api-access-z4pbx\") pod \"d6228ec7-9ffd-446c-b51c-99414b57969d\" (UID: \"d6228ec7-9ffd-446c-b51c-99414b57969d\") " Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.829229 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84edc32e-99b0-4d11-8aff-b445d6a03786-operator-scripts\") pod \"84edc32e-99b0-4d11-8aff-b445d6a03786\" (UID: \"84edc32e-99b0-4d11-8aff-b445d6a03786\") " Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.829301 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6228ec7-9ffd-446c-b51c-99414b57969d-operator-scripts\") pod \"d6228ec7-9ffd-446c-b51c-99414b57969d\" (UID: \"d6228ec7-9ffd-446c-b51c-99414b57969d\") " Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.829406 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz7p5\" (UniqueName: \"kubernetes.io/projected/84edc32e-99b0-4d11-8aff-b445d6a03786-kube-api-access-zz7p5\") pod \"84edc32e-99b0-4d11-8aff-b445d6a03786\" (UID: \"84edc32e-99b0-4d11-8aff-b445d6a03786\") " Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.830002 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84edc32e-99b0-4d11-8aff-b445d6a03786-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84edc32e-99b0-4d11-8aff-b445d6a03786" (UID: "84edc32e-99b0-4d11-8aff-b445d6a03786"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.830058 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6228ec7-9ffd-446c-b51c-99414b57969d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d6228ec7-9ffd-446c-b51c-99414b57969d" (UID: "d6228ec7-9ffd-446c-b51c-99414b57969d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.830345 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84edc32e-99b0-4d11-8aff-b445d6a03786-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.830549 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6228ec7-9ffd-446c-b51c-99414b57969d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.840422 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84edc32e-99b0-4d11-8aff-b445d6a03786-kube-api-access-zz7p5" (OuterVolumeSpecName: "kube-api-access-zz7p5") pod "84edc32e-99b0-4d11-8aff-b445d6a03786" (UID: "84edc32e-99b0-4d11-8aff-b445d6a03786"). InnerVolumeSpecName "kube-api-access-zz7p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.850685 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6228ec7-9ffd-446c-b51c-99414b57969d-kube-api-access-z4pbx" (OuterVolumeSpecName: "kube-api-access-z4pbx") pod "d6228ec7-9ffd-446c-b51c-99414b57969d" (UID: "d6228ec7-9ffd-446c-b51c-99414b57969d"). InnerVolumeSpecName "kube-api-access-z4pbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.931742 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz7p5\" (UniqueName: \"kubernetes.io/projected/84edc32e-99b0-4d11-8aff-b445d6a03786-kube-api-access-zz7p5\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:39 crc kubenswrapper[4931]: I1125 22:09:39.931775 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4pbx\" (UniqueName: \"kubernetes.io/projected/d6228ec7-9ffd-446c-b51c-99414b57969d-kube-api-access-z4pbx\") on node \"crc\" DevicePath \"\"" Nov 25 22:09:40 crc kubenswrapper[4931]: I1125 22:09:40.349530 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-x8gms" Nov 25 22:09:40 crc kubenswrapper[4931]: I1125 22:09:40.352016 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-x8gms" event={"ID":"84edc32e-99b0-4d11-8aff-b445d6a03786","Type":"ContainerDied","Data":"18d116ac29f7c6e3ca3c819fdd76a47bdeee66f080017dce21e5174de0e6ff4b"} Nov 25 22:09:40 crc kubenswrapper[4931]: I1125 22:09:40.352078 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18d116ac29f7c6e3ca3c819fdd76a47bdeee66f080017dce21e5174de0e6ff4b" Nov 25 22:09:40 crc kubenswrapper[4931]: I1125 22:09:40.354213 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" event={"ID":"d6228ec7-9ffd-446c-b51c-99414b57969d","Type":"ContainerDied","Data":"438d4313f87b37c7d1640002c8f08b8c4ce863b2db54abdd15d9bd92bb37777f"} Nov 25 22:09:40 crc kubenswrapper[4931]: I1125 22:09:40.354253 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="438d4313f87b37c7d1640002c8f08b8c4ce863b2db54abdd15d9bd92bb37777f" Nov 25 22:09:40 crc kubenswrapper[4931]: I1125 22:09:40.354299 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-311e-account-create-update-f5k5q" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.855081 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-fg6mp"] Nov 25 22:09:41 crc kubenswrapper[4931]: E1125 22:09:41.856473 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84edc32e-99b0-4d11-8aff-b445d6a03786" containerName="mariadb-database-create" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.856521 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="84edc32e-99b0-4d11-8aff-b445d6a03786" containerName="mariadb-database-create" Nov 25 22:09:41 crc kubenswrapper[4931]: E1125 22:09:41.856567 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6228ec7-9ffd-446c-b51c-99414b57969d" containerName="mariadb-account-create-update" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.856575 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6228ec7-9ffd-446c-b51c-99414b57969d" containerName="mariadb-account-create-update" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.856715 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="84edc32e-99b0-4d11-8aff-b445d6a03786" containerName="mariadb-database-create" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.856725 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6228ec7-9ffd-446c-b51c-99414b57969d" containerName="mariadb-account-create-update" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.860321 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.863410 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-d4qqf" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.863934 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-fg6mp"] Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.869337 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.870125 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-db-sync-config-data\") pod \"glance-db-sync-fg6mp\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.870180 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5t5c\" (UniqueName: \"kubernetes.io/projected/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-kube-api-access-q5t5c\") pod \"glance-db-sync-fg6mp\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.870250 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-config-data\") pod \"glance-db-sync-fg6mp\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.971572 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-db-sync-config-data\") pod \"glance-db-sync-fg6mp\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.971646 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5t5c\" (UniqueName: \"kubernetes.io/projected/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-kube-api-access-q5t5c\") pod \"glance-db-sync-fg6mp\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.972027 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-config-data\") pod \"glance-db-sync-fg6mp\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.976865 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-config-data\") pod \"glance-db-sync-fg6mp\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.983331 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-db-sync-config-data\") pod \"glance-db-sync-fg6mp\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:41 crc kubenswrapper[4931]: I1125 22:09:41.994386 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5t5c\" (UniqueName: \"kubernetes.io/projected/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-kube-api-access-q5t5c\") pod \"glance-db-sync-fg6mp\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:42 crc kubenswrapper[4931]: I1125 22:09:42.226996 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:09:46 crc kubenswrapper[4931]: I1125 22:09:46.236757 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-fg6mp"] Nov 25 22:09:46 crc kubenswrapper[4931]: W1125 22:09:46.247393 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3b99ea6_5d2b_4d1b_afbe_d49805c16ed1.slice/crio-9883369e1b0afb83a402ec71d3e745098d9f8e3df182698ca44b4451ffa676a1 WatchSource:0}: Error finding container 9883369e1b0afb83a402ec71d3e745098d9f8e3df182698ca44b4451ffa676a1: Status 404 returned error can't find the container with id 9883369e1b0afb83a402ec71d3e745098d9f8e3df182698ca44b4451ffa676a1 Nov 25 22:09:46 crc kubenswrapper[4931]: I1125 22:09:46.410699 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"e5f137dc-0889-45c9-bfb7-f99227ae93df","Type":"ContainerStarted","Data":"75b4fee7ba6c4bddf6b4927bb752fd09de086a31153c0e3295ebc8dadaf06320"} Nov 25 22:09:46 crc kubenswrapper[4931]: I1125 22:09:46.412449 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-fg6mp" event={"ID":"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1","Type":"ContainerStarted","Data":"9883369e1b0afb83a402ec71d3e745098d9f8e3df182698ca44b4451ffa676a1"} Nov 25 22:09:46 crc kubenswrapper[4931]: I1125 22:09:46.437626 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.007421797 podStartE2EDuration="10.437600773s" podCreationTimestamp="2025-11-25 22:09:36 +0000 UTC" firstStartedPulling="2025-11-25 22:09:37.565423202 +0000 UTC m=+1115.921935321" lastFinishedPulling="2025-11-25 22:09:45.995602178 +0000 UTC m=+1124.352114297" observedRunningTime="2025-11-25 22:09:46.433183615 +0000 UTC m=+1124.789695734" watchObservedRunningTime="2025-11-25 22:09:46.437600773 +0000 UTC m=+1124.794112922" Nov 25 22:09:57 crc kubenswrapper[4931]: I1125 22:09:57.508929 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-fg6mp" event={"ID":"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1","Type":"ContainerStarted","Data":"956b3f180c7078f5ab4378ba41a8c5b80cb6bb0cd768ddbc26f371ae0260b20b"} Nov 25 22:09:57 crc kubenswrapper[4931]: I1125 22:09:57.541426 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-fg6mp" podStartSLOduration=6.499511644 podStartE2EDuration="16.541408384s" podCreationTimestamp="2025-11-25 22:09:41 +0000 UTC" firstStartedPulling="2025-11-25 22:09:46.249667206 +0000 UTC m=+1124.606179325" lastFinishedPulling="2025-11-25 22:09:56.291563916 +0000 UTC m=+1134.648076065" observedRunningTime="2025-11-25 22:09:57.535696681 +0000 UTC m=+1135.892208840" watchObservedRunningTime="2025-11-25 22:09:57.541408384 +0000 UTC m=+1135.897920503" Nov 25 22:10:03 crc kubenswrapper[4931]: I1125 22:10:03.571146 4931 generic.go:334] "Generic (PLEG): container finished" podID="a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1" containerID="956b3f180c7078f5ab4378ba41a8c5b80cb6bb0cd768ddbc26f371ae0260b20b" exitCode=0 Nov 25 22:10:03 crc kubenswrapper[4931]: I1125 22:10:03.571254 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-fg6mp" event={"ID":"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1","Type":"ContainerDied","Data":"956b3f180c7078f5ab4378ba41a8c5b80cb6bb0cd768ddbc26f371ae0260b20b"} Nov 25 22:10:04 crc kubenswrapper[4931]: I1125 22:10:04.975297 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.144684 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-db-sync-config-data\") pod \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.144734 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-config-data\") pod \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.144963 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5t5c\" (UniqueName: \"kubernetes.io/projected/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-kube-api-access-q5t5c\") pod \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\" (UID: \"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1\") " Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.151367 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-kube-api-access-q5t5c" (OuterVolumeSpecName: "kube-api-access-q5t5c") pod "a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1" (UID: "a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1"). InnerVolumeSpecName "kube-api-access-q5t5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.151990 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1" (UID: "a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.196054 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-config-data" (OuterVolumeSpecName: "config-data") pod "a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1" (UID: "a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.247070 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5t5c\" (UniqueName: \"kubernetes.io/projected/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-kube-api-access-q5t5c\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.247131 4931 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.247155 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.595026 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-fg6mp" event={"ID":"a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1","Type":"ContainerDied","Data":"9883369e1b0afb83a402ec71d3e745098d9f8e3df182698ca44b4451ffa676a1"} Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.595114 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9883369e1b0afb83a402ec71d3e745098d9f8e3df182698ca44b4451ffa676a1" Nov 25 22:10:05 crc kubenswrapper[4931]: I1125 22:10:05.595122 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-fg6mp" Nov 25 22:10:06 crc kubenswrapper[4931]: I1125 22:10:06.986016 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 22:10:06 crc kubenswrapper[4931]: E1125 22:10:06.986636 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1" containerName="glance-db-sync" Nov 25 22:10:06 crc kubenswrapper[4931]: I1125 22:10:06.986654 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1" containerName="glance-db-sync" Nov 25 22:10:06 crc kubenswrapper[4931]: I1125 22:10:06.989902 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1" containerName="glance-db-sync" Nov 25 22:10:06 crc kubenswrapper[4931]: I1125 22:10:06.991631 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:06 crc kubenswrapper[4931]: I1125 22:10:06.998554 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-d4qqf" Nov 25 22:10:06 crc kubenswrapper[4931]: I1125 22:10:06.998588 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 22:10:06 crc kubenswrapper[4931]: I1125 22:10:06.998709 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.027301 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.059262 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.061686 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.102085 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.171986 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-config-data\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172034 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172060 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-nvme\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172110 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172131 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-logs\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172146 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-httpd-run\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172162 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172178 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172196 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-scripts\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172216 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-lib-modules\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172240 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172254 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-dev\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172278 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172292 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-logs\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172306 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-sys\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172326 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172340 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-scripts\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172355 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-sys\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172370 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172384 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-run\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172406 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp5jf\" (UniqueName: \"kubernetes.io/projected/508fe869-c56d-4599-b022-985955029c84-kube-api-access-cp5jf\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172420 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-config-data\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172435 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxfp6\" (UniqueName: \"kubernetes.io/projected/9bc0b077-1d87-4f17-a0df-2bda31bca040-kube-api-access-hxfp6\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172452 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172465 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-dev\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172480 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-httpd-run\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172499 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-run\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.172516 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-lib-modules\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.273614 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-sys\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.273925 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274014 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-sys\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274110 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-scripts\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274152 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.273773 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-sys\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274216 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-sys\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274220 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274326 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-run\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274366 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp5jf\" (UniqueName: \"kubernetes.io/projected/508fe869-c56d-4599-b022-985955029c84-kube-api-access-cp5jf\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274392 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-config-data\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274415 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxfp6\" (UniqueName: \"kubernetes.io/projected/9bc0b077-1d87-4f17-a0df-2bda31bca040-kube-api-access-hxfp6\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274443 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274462 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-dev\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274488 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-httpd-run\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274514 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-run\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274538 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-lib-modules\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274571 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-config-data\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274605 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274639 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-nvme\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274678 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274705 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-logs\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274731 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274751 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-httpd-run\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274780 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274789 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-dev\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.275171 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.275191 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-run\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.275341 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-httpd-run\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.275393 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-run\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.275438 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-lib-modules\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.275805 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.275969 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-logs\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276218 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276282 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-nvme\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276315 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.274808 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-scripts\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276370 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-lib-modules\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276419 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276450 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-dev\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276492 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276514 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-logs\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276836 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-logs\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276902 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-lib-modules\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.276948 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.277014 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.277055 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.277072 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.277154 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-dev\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.277567 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-httpd-run\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.279748 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-scripts\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.283927 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-scripts\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.285046 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-config-data\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.296423 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.307171 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-config-data\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.309691 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp5jf\" (UniqueName: \"kubernetes.io/projected/508fe869-c56d-4599-b022-985955029c84-kube-api-access-cp5jf\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.310361 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxfp6\" (UniqueName: \"kubernetes.io/projected/9bc0b077-1d87-4f17-a0df-2bda31bca040-kube-api-access-hxfp6\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.310960 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.324715 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.330466 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.377900 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.626593 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:07 crc kubenswrapper[4931]: I1125 22:10:07.753299 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:07 crc kubenswrapper[4931]: W1125 22:10:07.759937 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9bc0b077_1d87_4f17_a0df_2bda31bca040.slice/crio-25ecb14aa0802a70d4ffe9e0cfae8dbb6c2d66a3c8547ee03c01ee444d6755a5 WatchSource:0}: Error finding container 25ecb14aa0802a70d4ffe9e0cfae8dbb6c2d66a3c8547ee03c01ee444d6755a5: Status 404 returned error can't find the container with id 25ecb14aa0802a70d4ffe9e0cfae8dbb6c2d66a3c8547ee03c01ee444d6755a5 Nov 25 22:10:08 crc kubenswrapper[4931]: I1125 22:10:08.045257 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 22:10:08 crc kubenswrapper[4931]: I1125 22:10:08.624140 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"508fe869-c56d-4599-b022-985955029c84","Type":"ContainerStarted","Data":"f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393"} Nov 25 22:10:08 crc kubenswrapper[4931]: I1125 22:10:08.624739 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"508fe869-c56d-4599-b022-985955029c84","Type":"ContainerStarted","Data":"1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea"} Nov 25 22:10:08 crc kubenswrapper[4931]: I1125 22:10:08.624757 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"508fe869-c56d-4599-b022-985955029c84","Type":"ContainerStarted","Data":"1a9a42663c7151bacc3422f2749d41ed10f842902660db11d85828441bdf8d2e"} Nov 25 22:10:08 crc kubenswrapper[4931]: I1125 22:10:08.628085 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9bc0b077-1d87-4f17-a0df-2bda31bca040","Type":"ContainerStarted","Data":"50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6"} Nov 25 22:10:08 crc kubenswrapper[4931]: I1125 22:10:08.628128 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9bc0b077-1d87-4f17-a0df-2bda31bca040","Type":"ContainerStarted","Data":"d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b"} Nov 25 22:10:08 crc kubenswrapper[4931]: I1125 22:10:08.628141 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9bc0b077-1d87-4f17-a0df-2bda31bca040","Type":"ContainerStarted","Data":"25ecb14aa0802a70d4ffe9e0cfae8dbb6c2d66a3c8547ee03c01ee444d6755a5"} Nov 25 22:10:08 crc kubenswrapper[4931]: I1125 22:10:08.649529 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.649186274 podStartE2EDuration="2.649186274s" podCreationTimestamp="2025-11-25 22:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:10:08.645009521 +0000 UTC m=+1147.001521630" watchObservedRunningTime="2025-11-25 22:10:08.649186274 +0000 UTC m=+1147.005698403" Nov 25 22:10:08 crc kubenswrapper[4931]: I1125 22:10:08.674642 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.674624328 podStartE2EDuration="2.674624328s" podCreationTimestamp="2025-11-25 22:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:10:08.670111947 +0000 UTC m=+1147.026624066" watchObservedRunningTime="2025-11-25 22:10:08.674624328 +0000 UTC m=+1147.031136447" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.379381 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.380099 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.416787 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.447950 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.628102 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.628159 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.672845 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.679290 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.709691 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.711559 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.711590 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:17 crc kubenswrapper[4931]: I1125 22:10:17.711603 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:19 crc kubenswrapper[4931]: I1125 22:10:19.721061 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 22:10:19 crc kubenswrapper[4931]: I1125 22:10:19.721319 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 22:10:19 crc kubenswrapper[4931]: I1125 22:10:19.722078 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 22:10:19 crc kubenswrapper[4931]: I1125 22:10:19.722090 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 22:10:19 crc kubenswrapper[4931]: I1125 22:10:19.881667 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:19 crc kubenswrapper[4931]: I1125 22:10:19.882246 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:19 crc kubenswrapper[4931]: I1125 22:10:19.965458 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:20 crc kubenswrapper[4931]: I1125 22:10:20.030430 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:10:20 crc kubenswrapper[4931]: I1125 22:10:20.088356 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:21 crc kubenswrapper[4931]: I1125 22:10:21.736290 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerName="glance-log" containerID="cri-o://d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b" gracePeriod=30 Nov 25 22:10:21 crc kubenswrapper[4931]: I1125 22:10:21.736749 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerName="glance-httpd" containerID="cri-o://50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6" gracePeriod=30 Nov 25 22:10:22 crc kubenswrapper[4931]: I1125 22:10:22.746387 4931 generic.go:334] "Generic (PLEG): container finished" podID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerID="d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b" exitCode=143 Nov 25 22:10:22 crc kubenswrapper[4931]: I1125 22:10:22.746438 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9bc0b077-1d87-4f17-a0df-2bda31bca040","Type":"ContainerDied","Data":"d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b"} Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.340552 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.418998 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-sys\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419072 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-logs\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419104 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-dev\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419157 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-run\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419190 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-config-data\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419262 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419315 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-nvme\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419388 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxfp6\" (UniqueName: \"kubernetes.io/projected/9bc0b077-1d87-4f17-a0df-2bda31bca040-kube-api-access-hxfp6\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419430 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-scripts\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419470 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-iscsi\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419606 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-var-locks-brick\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419673 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-lib-modules\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419720 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.419754 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-httpd-run\") pod \"9bc0b077-1d87-4f17-a0df-2bda31bca040\" (UID: \"9bc0b077-1d87-4f17-a0df-2bda31bca040\") " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.420678 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.420745 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-sys" (OuterVolumeSpecName: "sys") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.421171 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-logs" (OuterVolumeSpecName: "logs") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.421219 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-dev" (OuterVolumeSpecName: "dev") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.421255 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-run" (OuterVolumeSpecName: "run") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.422290 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.422393 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.422467 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.422847 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.437666 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bc0b077-1d87-4f17-a0df-2bda31bca040-kube-api-access-hxfp6" (OuterVolumeSpecName: "kube-api-access-hxfp6") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "kube-api-access-hxfp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.439089 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.439133 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-scripts" (OuterVolumeSpecName: "scripts") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.445406 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.479700 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-config-data" (OuterVolumeSpecName: "config-data") pod "9bc0b077-1d87-4f17-a0df-2bda31bca040" (UID: "9bc0b077-1d87-4f17-a0df-2bda31bca040"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522185 4931 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522241 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxfp6\" (UniqueName: \"kubernetes.io/projected/9bc0b077-1d87-4f17-a0df-2bda31bca040-kube-api-access-hxfp6\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522262 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522279 4931 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522296 4931 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522312 4931 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522345 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522361 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522378 4931 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-sys\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522393 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bc0b077-1d87-4f17-a0df-2bda31bca040-logs\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522409 4931 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-dev\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522422 4931 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9bc0b077-1d87-4f17-a0df-2bda31bca040-run\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522436 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bc0b077-1d87-4f17-a0df-2bda31bca040-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.522459 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.540036 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.547178 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.624336 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.624384 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.773925 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.773932 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9bc0b077-1d87-4f17-a0df-2bda31bca040","Type":"ContainerDied","Data":"50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6"} Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.774054 4931 scope.go:117] "RemoveContainer" containerID="50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.773863 4931 generic.go:334] "Generic (PLEG): container finished" podID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerID="50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6" exitCode=0 Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.774157 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9bc0b077-1d87-4f17-a0df-2bda31bca040","Type":"ContainerDied","Data":"25ecb14aa0802a70d4ffe9e0cfae8dbb6c2d66a3c8547ee03c01ee444d6755a5"} Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.802134 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.804461 4931 scope.go:117] "RemoveContainer" containerID="d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.811557 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.825821 4931 scope.go:117] "RemoveContainer" containerID="50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6" Nov 25 22:10:25 crc kubenswrapper[4931]: E1125 22:10:25.828310 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6\": container with ID starting with 50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6 not found: ID does not exist" containerID="50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.828340 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6"} err="failed to get container status \"50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6\": rpc error: code = NotFound desc = could not find container \"50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6\": container with ID starting with 50540e4c18f7ede0b7a12ee6e330dc79480a79c6a2d21dbbbd6db8ad4af5b4b6 not found: ID does not exist" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.828358 4931 scope.go:117] "RemoveContainer" containerID="d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b" Nov 25 22:10:25 crc kubenswrapper[4931]: E1125 22:10:25.828706 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b\": container with ID starting with d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b not found: ID does not exist" containerID="d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.828838 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b"} err="failed to get container status \"d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b\": rpc error: code = NotFound desc = could not find container \"d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b\": container with ID starting with d3172d9bc2e7d805e0539e8519e94e70812afac5dcc82d25f05b8dc07e82d12b not found: ID does not exist" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.833268 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:25 crc kubenswrapper[4931]: E1125 22:10:25.833589 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerName="glance-log" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.833611 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerName="glance-log" Nov 25 22:10:25 crc kubenswrapper[4931]: E1125 22:10:25.833645 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerName="glance-httpd" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.833653 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerName="glance-httpd" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.833796 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerName="glance-log" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.833822 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bc0b077-1d87-4f17-a0df-2bda31bca040" containerName="glance-httpd" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.835380 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.851296 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928184 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlpkq\" (UniqueName: \"kubernetes.io/projected/9c41a041-45ce-491b-a319-f28895487aec-kube-api-access-mlpkq\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928230 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-run\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928263 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928288 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928303 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928321 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-httpd-run\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928340 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-config-data\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928359 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-sys\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928379 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-dev\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928408 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-scripts\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928430 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928465 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-logs\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928499 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:25 crc kubenswrapper[4931]: I1125 22:10:25.928527 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-lib-modules\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030009 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-lib-modules\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030412 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlpkq\" (UniqueName: \"kubernetes.io/projected/9c41a041-45ce-491b-a319-f28895487aec-kube-api-access-mlpkq\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030114 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-lib-modules\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030459 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-run\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030511 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-run\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030621 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030695 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030722 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030757 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-httpd-run\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030796 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-config-data\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030837 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-sys\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030892 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-dev\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030985 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-scripts\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031027 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031045 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031100 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-logs\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031201 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031235 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031324 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-sys\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031400 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-dev\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031471 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.030986 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031692 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.031710 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-httpd-run\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.032185 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-logs\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.037018 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-scripts\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.045725 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-config-data\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.046945 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlpkq\" (UniqueName: \"kubernetes.io/projected/9c41a041-45ce-491b-a319-f28895487aec-kube-api-access-mlpkq\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.056331 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.067181 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.149994 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.432358 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bc0b077-1d87-4f17-a0df-2bda31bca040" path="/var/lib/kubelet/pods/9bc0b077-1d87-4f17-a0df-2bda31bca040/volumes" Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.663199 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:26 crc kubenswrapper[4931]: I1125 22:10:26.784808 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9c41a041-45ce-491b-a319-f28895487aec","Type":"ContainerStarted","Data":"ec5cf2b92ac4dd73399f2a1d5f5633a8f938b75a4dd83422d76cbea5bbaa6db3"} Nov 25 22:10:27 crc kubenswrapper[4931]: I1125 22:10:27.798796 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9c41a041-45ce-491b-a319-f28895487aec","Type":"ContainerStarted","Data":"ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be"} Nov 25 22:10:27 crc kubenswrapper[4931]: I1125 22:10:27.799475 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9c41a041-45ce-491b-a319-f28895487aec","Type":"ContainerStarted","Data":"3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e"} Nov 25 22:10:27 crc kubenswrapper[4931]: I1125 22:10:27.829202 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.829176197 podStartE2EDuration="2.829176197s" podCreationTimestamp="2025-11-25 22:10:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:10:27.824723737 +0000 UTC m=+1166.181235896" watchObservedRunningTime="2025-11-25 22:10:27.829176197 +0000 UTC m=+1166.185688346" Nov 25 22:10:36 crc kubenswrapper[4931]: I1125 22:10:36.150384 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:36 crc kubenswrapper[4931]: I1125 22:10:36.151143 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:36 crc kubenswrapper[4931]: I1125 22:10:36.190558 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:36 crc kubenswrapper[4931]: I1125 22:10:36.219142 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:36 crc kubenswrapper[4931]: I1125 22:10:36.869988 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:36 crc kubenswrapper[4931]: I1125 22:10:36.870019 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:38 crc kubenswrapper[4931]: I1125 22:10:38.860325 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:38 crc kubenswrapper[4931]: I1125 22:10:38.883956 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 22:10:38 crc kubenswrapper[4931]: I1125 22:10:38.947183 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.460953 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-fg6mp"] Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.469427 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-fg6mp"] Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.568392 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-shcq9"] Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.569639 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.578588 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-shcq9"] Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.579131 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.579357 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.659356 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-combined-ca-bundle\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.659413 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-config-data\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.659661 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-db-sync-config-data\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.659728 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvsdm\" (UniqueName: \"kubernetes.io/projected/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-kube-api-access-mvsdm\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.761702 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvsdm\" (UniqueName: \"kubernetes.io/projected/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-kube-api-access-mvsdm\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.762339 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-combined-ca-bundle\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.762581 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-config-data\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.763044 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-db-sync-config-data\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.775832 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-db-sync-config-data\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.776222 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-combined-ca-bundle\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.776536 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-config-data\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.782756 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvsdm\" (UniqueName: \"kubernetes.io/projected/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-kube-api-access-mvsdm\") pod \"glance-db-sync-shcq9\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:52 crc kubenswrapper[4931]: I1125 22:10:52.896925 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:53 crc kubenswrapper[4931]: I1125 22:10:53.337550 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-shcq9"] Nov 25 22:10:54 crc kubenswrapper[4931]: I1125 22:10:54.012450 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-shcq9" event={"ID":"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8","Type":"ContainerStarted","Data":"85e8b871e841792add377f6341603951c2f6231fb5285d682bae046f7e4539e2"} Nov 25 22:10:54 crc kubenswrapper[4931]: I1125 22:10:54.012781 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-shcq9" event={"ID":"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8","Type":"ContainerStarted","Data":"e82ca7e70c7281e3493176da3829d16d61d22d81f0d6976342e155c73742e03d"} Nov 25 22:10:54 crc kubenswrapper[4931]: I1125 22:10:54.040409 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-shcq9" podStartSLOduration=2.040389053 podStartE2EDuration="2.040389053s" podCreationTimestamp="2025-11-25 22:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:10:54.03689021 +0000 UTC m=+1192.393402349" watchObservedRunningTime="2025-11-25 22:10:54.040389053 +0000 UTC m=+1192.396901172" Nov 25 22:10:54 crc kubenswrapper[4931]: I1125 22:10:54.436201 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1" path="/var/lib/kubelet/pods/a3b99ea6-5d2b-4d1b-afbe-d49805c16ed1/volumes" Nov 25 22:10:57 crc kubenswrapper[4931]: I1125 22:10:57.040970 4931 generic.go:334] "Generic (PLEG): container finished" podID="fa092118-ff34-40e3-8f23-bcc5dc2ae1d8" containerID="85e8b871e841792add377f6341603951c2f6231fb5285d682bae046f7e4539e2" exitCode=0 Nov 25 22:10:57 crc kubenswrapper[4931]: I1125 22:10:57.041027 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-shcq9" event={"ID":"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8","Type":"ContainerDied","Data":"85e8b871e841792add377f6341603951c2f6231fb5285d682bae046f7e4539e2"} Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.381199 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.564234 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvsdm\" (UniqueName: \"kubernetes.io/projected/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-kube-api-access-mvsdm\") pod \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.564330 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-combined-ca-bundle\") pod \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.564512 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-config-data\") pod \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.564544 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-db-sync-config-data\") pod \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\" (UID: \"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8\") " Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.571092 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fa092118-ff34-40e3-8f23-bcc5dc2ae1d8" (UID: "fa092118-ff34-40e3-8f23-bcc5dc2ae1d8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.571743 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-kube-api-access-mvsdm" (OuterVolumeSpecName: "kube-api-access-mvsdm") pod "fa092118-ff34-40e3-8f23-bcc5dc2ae1d8" (UID: "fa092118-ff34-40e3-8f23-bcc5dc2ae1d8"). InnerVolumeSpecName "kube-api-access-mvsdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.600049 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa092118-ff34-40e3-8f23-bcc5dc2ae1d8" (UID: "fa092118-ff34-40e3-8f23-bcc5dc2ae1d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.607153 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-config-data" (OuterVolumeSpecName: "config-data") pod "fa092118-ff34-40e3-8f23-bcc5dc2ae1d8" (UID: "fa092118-ff34-40e3-8f23-bcc5dc2ae1d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.666144 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.666203 4931 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.666226 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvsdm\" (UniqueName: \"kubernetes.io/projected/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-kube-api-access-mvsdm\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:58 crc kubenswrapper[4931]: I1125 22:10:58.666245 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:10:59 crc kubenswrapper[4931]: I1125 22:10:59.058651 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-shcq9" event={"ID":"fa092118-ff34-40e3-8f23-bcc5dc2ae1d8","Type":"ContainerDied","Data":"e82ca7e70c7281e3493176da3829d16d61d22d81f0d6976342e155c73742e03d"} Nov 25 22:10:59 crc kubenswrapper[4931]: I1125 22:10:59.058696 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e82ca7e70c7281e3493176da3829d16d61d22d81f0d6976342e155c73742e03d" Nov 25 22:10:59 crc kubenswrapper[4931]: I1125 22:10:59.058693 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-shcq9" Nov 25 22:10:59 crc kubenswrapper[4931]: I1125 22:10:59.245720 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:10:59 crc kubenswrapper[4931]: I1125 22:10:59.246633 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9c41a041-45ce-491b-a319-f28895487aec" containerName="glance-log" containerID="cri-o://3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e" gracePeriod=30 Nov 25 22:10:59 crc kubenswrapper[4931]: I1125 22:10:59.246843 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9c41a041-45ce-491b-a319-f28895487aec" containerName="glance-httpd" containerID="cri-o://ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be" gracePeriod=30 Nov 25 22:10:59 crc kubenswrapper[4931]: I1125 22:10:59.279102 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 22:10:59 crc kubenswrapper[4931]: I1125 22:10:59.279538 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="508fe869-c56d-4599-b022-985955029c84" containerName="glance-log" containerID="cri-o://1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea" gracePeriod=30 Nov 25 22:10:59 crc kubenswrapper[4931]: I1125 22:10:59.280202 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="508fe869-c56d-4599-b022-985955029c84" containerName="glance-httpd" containerID="cri-o://f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393" gracePeriod=30 Nov 25 22:11:00 crc kubenswrapper[4931]: I1125 22:11:00.074808 4931 generic.go:334] "Generic (PLEG): container finished" podID="508fe869-c56d-4599-b022-985955029c84" containerID="1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea" exitCode=143 Nov 25 22:11:00 crc kubenswrapper[4931]: I1125 22:11:00.075953 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"508fe869-c56d-4599-b022-985955029c84","Type":"ContainerDied","Data":"1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea"} Nov 25 22:11:00 crc kubenswrapper[4931]: I1125 22:11:00.077766 4931 generic.go:334] "Generic (PLEG): container finished" podID="9c41a041-45ce-491b-a319-f28895487aec" containerID="3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e" exitCode=143 Nov 25 22:11:00 crc kubenswrapper[4931]: I1125 22:11:00.077802 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9c41a041-45ce-491b-a319-f28895487aec","Type":"ContainerDied","Data":"3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e"} Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.892906 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.897650 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941354 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-config-data\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941610 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-logs\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941642 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-config-data\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941659 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941680 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941715 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-nvme\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941733 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-var-locks-brick\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941751 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-dev\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941772 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-scripts\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941805 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941827 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-run\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941851 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-lib-modules\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941869 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-lib-modules\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941905 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-sys\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941928 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-run\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941947 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-httpd-run\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.941974 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-scripts\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942005 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-dev\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942059 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942080 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-dev" (OuterVolumeSpecName: "dev") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942522 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-sys\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942552 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp5jf\" (UniqueName: \"kubernetes.io/projected/508fe869-c56d-4599-b022-985955029c84-kube-api-access-cp5jf\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942586 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlpkq\" (UniqueName: \"kubernetes.io/projected/9c41a041-45ce-491b-a319-f28895487aec-kube-api-access-mlpkq\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942604 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-nvme\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942622 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-iscsi\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942641 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"9c41a041-45ce-491b-a319-f28895487aec\" (UID: \"9c41a041-45ce-491b-a319-f28895487aec\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942671 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-iscsi\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942692 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-httpd-run\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942722 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-var-locks-brick\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942746 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-logs\") pod \"508fe869-c56d-4599-b022-985955029c84\" (UID: \"508fe869-c56d-4599-b022-985955029c84\") " Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.942839 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-logs" (OuterVolumeSpecName: "logs") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.943067 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-logs\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.943080 4931 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-dev\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.943090 4931 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.943464 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-logs" (OuterVolumeSpecName: "logs") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.943496 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-sys" (OuterVolumeSpecName: "sys") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.943518 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-run" (OuterVolumeSpecName: "run") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.943741 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.948606 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.948669 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-dev" (OuterVolumeSpecName: "dev") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.948675 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.948700 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-sys" (OuterVolumeSpecName: "sys") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.948717 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-run" (OuterVolumeSpecName: "run") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.948812 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-scripts" (OuterVolumeSpecName: "scripts") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.948845 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.949146 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.949185 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.949210 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.951637 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.956789 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508fe869-c56d-4599-b022-985955029c84-kube-api-access-cp5jf" (OuterVolumeSpecName: "kube-api-access-cp5jf") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "kube-api-access-cp5jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.957670 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.974238 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.974387 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-scripts" (OuterVolumeSpecName: "scripts") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.974534 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c41a041-45ce-491b-a319-f28895487aec-kube-api-access-mlpkq" (OuterVolumeSpecName: "kube-api-access-mlpkq") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "kube-api-access-mlpkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.974901 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.975031 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 22:11:02 crc kubenswrapper[4931]: I1125 22:11:02.977037 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.009325 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-config-data" (OuterVolumeSpecName: "config-data") pod "9c41a041-45ce-491b-a319-f28895487aec" (UID: "9c41a041-45ce-491b-a319-f28895487aec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.012634 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-config-data" (OuterVolumeSpecName: "config-data") pod "508fe869-c56d-4599-b022-985955029c84" (UID: "508fe869-c56d-4599-b022-985955029c84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.044336 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.044472 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.044515 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.044557 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045080 4931 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045103 4931 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045119 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c41a041-45ce-491b-a319-f28895487aec-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045148 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045161 4931 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-run\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045174 4931 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045185 4931 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-run\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045198 4931 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-sys\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045208 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c41a041-45ce-491b-a319-f28895487aec-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045218 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508fe869-c56d-4599-b022-985955029c84-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045230 4931 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-dev\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045241 4931 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-sys\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045252 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp5jf\" (UniqueName: \"kubernetes.io/projected/508fe869-c56d-4599-b022-985955029c84-kube-api-access-cp5jf\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045265 4931 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045276 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlpkq\" (UniqueName: \"kubernetes.io/projected/9c41a041-45ce-491b-a319-f28895487aec-kube-api-access-mlpkq\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045286 4931 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9c41a041-45ce-491b-a319-f28895487aec-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045311 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045323 4931 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045335 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045345 4931 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/508fe869-c56d-4599-b022-985955029c84-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.045356 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508fe869-c56d-4599-b022-985955029c84-logs\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.061522 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.062803 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.064479 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.067330 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.100967 4931 generic.go:334] "Generic (PLEG): container finished" podID="508fe869-c56d-4599-b022-985955029c84" containerID="f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393" exitCode=0 Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.101031 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"508fe869-c56d-4599-b022-985955029c84","Type":"ContainerDied","Data":"f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393"} Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.101060 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"508fe869-c56d-4599-b022-985955029c84","Type":"ContainerDied","Data":"1a9a42663c7151bacc3422f2749d41ed10f842902660db11d85828441bdf8d2e"} Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.101081 4931 scope.go:117] "RemoveContainer" containerID="f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.101089 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.105911 4931 generic.go:334] "Generic (PLEG): container finished" podID="9c41a041-45ce-491b-a319-f28895487aec" containerID="ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be" exitCode=0 Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.105939 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9c41a041-45ce-491b-a319-f28895487aec","Type":"ContainerDied","Data":"ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be"} Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.105959 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9c41a041-45ce-491b-a319-f28895487aec","Type":"ContainerDied","Data":"ec5cf2b92ac4dd73399f2a1d5f5633a8f938b75a4dd83422d76cbea5bbaa6db3"} Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.106019 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.119178 4931 scope.go:117] "RemoveContainer" containerID="1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.136188 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.137029 4931 scope.go:117] "RemoveContainer" containerID="f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393" Nov 25 22:11:03 crc kubenswrapper[4931]: E1125 22:11:03.138323 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393\": container with ID starting with f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393 not found: ID does not exist" containerID="f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.138358 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393"} err="failed to get container status \"f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393\": rpc error: code = NotFound desc = could not find container \"f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393\": container with ID starting with f130cd8d08461403ed3607db34f0337a11bc429bd13ddb729f6ad2312f918393 not found: ID does not exist" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.138383 4931 scope.go:117] "RemoveContainer" containerID="1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea" Nov 25 22:11:03 crc kubenswrapper[4931]: E1125 22:11:03.144423 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea\": container with ID starting with 1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea not found: ID does not exist" containerID="1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.144466 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea"} err="failed to get container status \"1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea\": rpc error: code = NotFound desc = could not find container \"1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea\": container with ID starting with 1f3ec2a0095400a637cf716799107e9ed4869c62b99130c0e55f79a6892dedea not found: ID does not exist" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.144496 4931 scope.go:117] "RemoveContainer" containerID="ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.146682 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.146711 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.146723 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.146736 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.150393 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.155793 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.162235 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.168433 4931 scope.go:117] "RemoveContainer" containerID="3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.185513 4931 scope.go:117] "RemoveContainer" containerID="ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be" Nov 25 22:11:03 crc kubenswrapper[4931]: E1125 22:11:03.185853 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be\": container with ID starting with ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be not found: ID does not exist" containerID="ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.185905 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be"} err="failed to get container status \"ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be\": rpc error: code = NotFound desc = could not find container \"ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be\": container with ID starting with ad3839b877572ad9f6ba06e2929689622b7b7dfc7ccebf1f9dd4058eb2e491be not found: ID does not exist" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.185933 4931 scope.go:117] "RemoveContainer" containerID="3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e" Nov 25 22:11:03 crc kubenswrapper[4931]: E1125 22:11:03.186179 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e\": container with ID starting with 3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e not found: ID does not exist" containerID="3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e" Nov 25 22:11:03 crc kubenswrapper[4931]: I1125 22:11:03.186201 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e"} err="failed to get container status \"3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e\": rpc error: code = NotFound desc = could not find container \"3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e\": container with ID starting with 3b8a6be3945f344ed99647f3b8356e1500c814e0e9b208c53657092195f2ee4e not found: ID does not exist" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.439209 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="508fe869-c56d-4599-b022-985955029c84" path="/var/lib/kubelet/pods/508fe869-c56d-4599-b022-985955029c84/volumes" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.440995 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c41a041-45ce-491b-a319-f28895487aec" path="/var/lib/kubelet/pods/9c41a041-45ce-491b-a319-f28895487aec/volumes" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.588665 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:11:04 crc kubenswrapper[4931]: E1125 22:11:04.589136 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c41a041-45ce-491b-a319-f28895487aec" containerName="glance-httpd" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589160 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c41a041-45ce-491b-a319-f28895487aec" containerName="glance-httpd" Nov 25 22:11:04 crc kubenswrapper[4931]: E1125 22:11:04.589184 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508fe869-c56d-4599-b022-985955029c84" containerName="glance-log" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589198 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="508fe869-c56d-4599-b022-985955029c84" containerName="glance-log" Nov 25 22:11:04 crc kubenswrapper[4931]: E1125 22:11:04.589228 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c41a041-45ce-491b-a319-f28895487aec" containerName="glance-log" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589240 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c41a041-45ce-491b-a319-f28895487aec" containerName="glance-log" Nov 25 22:11:04 crc kubenswrapper[4931]: E1125 22:11:04.589262 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa092118-ff34-40e3-8f23-bcc5dc2ae1d8" containerName="glance-db-sync" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589274 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa092118-ff34-40e3-8f23-bcc5dc2ae1d8" containerName="glance-db-sync" Nov 25 22:11:04 crc kubenswrapper[4931]: E1125 22:11:04.589290 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508fe869-c56d-4599-b022-985955029c84" containerName="glance-httpd" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589303 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="508fe869-c56d-4599-b022-985955029c84" containerName="glance-httpd" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589521 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="508fe869-c56d-4599-b022-985955029c84" containerName="glance-httpd" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589549 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa092118-ff34-40e3-8f23-bcc5dc2ae1d8" containerName="glance-db-sync" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589566 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="508fe869-c56d-4599-b022-985955029c84" containerName="glance-log" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589596 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c41a041-45ce-491b-a319-f28895487aec" containerName="glance-httpd" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.589622 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c41a041-45ce-491b-a319-f28895487aec" containerName="glance-log" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.591082 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.601140 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.601347 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.601381 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.601390 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.602058 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-d4qqf" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.622340 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.634336 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.670565 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.670648 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.670699 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-scripts\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.670748 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.670799 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-config-data\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.670826 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.670921 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-logs\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.670960 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-httpd-run\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.671089 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgvkc\" (UniqueName: \"kubernetes.io/projected/338c52e4-b31c-48df-8a37-a1a36c9fadf9-kube-api-access-jgvkc\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.771828 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgvkc\" (UniqueName: \"kubernetes.io/projected/338c52e4-b31c-48df-8a37-a1a36c9fadf9-kube-api-access-jgvkc\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.772080 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.772225 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.772379 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-scripts\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.772562 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.772713 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-config-data\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.772836 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.773043 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-logs\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.773175 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-httpd-run\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.773498 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-logs\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.772931 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.773929 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-httpd-run\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.777811 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.778203 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-scripts\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.781788 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-config-data\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.782032 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.783361 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.798627 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgvkc\" (UniqueName: \"kubernetes.io/projected/338c52e4-b31c-48df-8a37-a1a36c9fadf9-kube-api-access-jgvkc\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.802572 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-0\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:04 crc kubenswrapper[4931]: I1125 22:11:04.947564 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:05 crc kubenswrapper[4931]: I1125 22:11:05.474978 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:11:06 crc kubenswrapper[4931]: I1125 22:11:06.138576 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"338c52e4-b31c-48df-8a37-a1a36c9fadf9","Type":"ContainerStarted","Data":"06c8d3fe04995a69d7556d1ea41b0d356fb01e45e5b10aa40e9b8af8f4ed07be"} Nov 25 22:11:07 crc kubenswrapper[4931]: I1125 22:11:07.147206 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"338c52e4-b31c-48df-8a37-a1a36c9fadf9","Type":"ContainerStarted","Data":"4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be"} Nov 25 22:11:07 crc kubenswrapper[4931]: I1125 22:11:07.147566 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"338c52e4-b31c-48df-8a37-a1a36c9fadf9","Type":"ContainerStarted","Data":"64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee"} Nov 25 22:11:07 crc kubenswrapper[4931]: I1125 22:11:07.172046 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.172018993 podStartE2EDuration="3.172018993s" podCreationTimestamp="2025-11-25 22:11:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:11:07.166451943 +0000 UTC m=+1205.522964072" watchObservedRunningTime="2025-11-25 22:11:07.172018993 +0000 UTC m=+1205.528531122" Nov 25 22:11:08 crc kubenswrapper[4931]: I1125 22:11:08.846296 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:11:08 crc kubenswrapper[4931]: I1125 22:11:08.846344 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:11:14 crc kubenswrapper[4931]: I1125 22:11:14.948624 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:14 crc kubenswrapper[4931]: I1125 22:11:14.949307 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:14 crc kubenswrapper[4931]: I1125 22:11:14.994262 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:15 crc kubenswrapper[4931]: I1125 22:11:15.017547 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:15 crc kubenswrapper[4931]: I1125 22:11:15.251218 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:15 crc kubenswrapper[4931]: I1125 22:11:15.251532 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:17 crc kubenswrapper[4931]: I1125 22:11:17.071342 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:17 crc kubenswrapper[4931]: I1125 22:11:17.133586 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.007479 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-shcq9"] Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.015314 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-shcq9"] Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.069058 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.100947 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance311e-account-delete-fq67b"] Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.104125 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.119069 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance311e-account-delete-fq67b"] Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.206593 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcslh\" (UniqueName: \"kubernetes.io/projected/31eac098-adf3-49d3-8bff-601cfc96c4a5-kube-api-access-kcslh\") pod \"glance311e-account-delete-fq67b\" (UID: \"31eac098-adf3-49d3-8bff-601cfc96c4a5\") " pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.206668 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eac098-adf3-49d3-8bff-601cfc96c4a5-operator-scripts\") pod \"glance311e-account-delete-fq67b\" (UID: \"31eac098-adf3-49d3-8bff-601cfc96c4a5\") " pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.278565 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerName="glance-log" containerID="cri-o://64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee" gracePeriod=30 Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.278661 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerName="glance-httpd" containerID="cri-o://4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be" gracePeriod=30 Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.282822 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.104:9292/healthcheck\": EOF" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.308489 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcslh\" (UniqueName: \"kubernetes.io/projected/31eac098-adf3-49d3-8bff-601cfc96c4a5-kube-api-access-kcslh\") pod \"glance311e-account-delete-fq67b\" (UID: \"31eac098-adf3-49d3-8bff-601cfc96c4a5\") " pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.308548 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eac098-adf3-49d3-8bff-601cfc96c4a5-operator-scripts\") pod \"glance311e-account-delete-fq67b\" (UID: \"31eac098-adf3-49d3-8bff-601cfc96c4a5\") " pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.309345 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eac098-adf3-49d3-8bff-601cfc96c4a5-operator-scripts\") pod \"glance311e-account-delete-fq67b\" (UID: \"31eac098-adf3-49d3-8bff-601cfc96c4a5\") " pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.333731 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcslh\" (UniqueName: \"kubernetes.io/projected/31eac098-adf3-49d3-8bff-601cfc96c4a5-kube-api-access-kcslh\") pod \"glance311e-account-delete-fq67b\" (UID: \"31eac098-adf3-49d3-8bff-601cfc96c4a5\") " pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.418734 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:19 crc kubenswrapper[4931]: I1125 22:11:19.874936 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance311e-account-delete-fq67b"] Nov 25 22:11:20 crc kubenswrapper[4931]: I1125 22:11:20.286318 4931 generic.go:334] "Generic (PLEG): container finished" podID="31eac098-adf3-49d3-8bff-601cfc96c4a5" containerID="3b241f4f0d61953fd8db9f03162f41c15f4640b20ed666996d54bbdd4f451966" exitCode=0 Nov 25 22:11:20 crc kubenswrapper[4931]: I1125 22:11:20.286412 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance311e-account-delete-fq67b" event={"ID":"31eac098-adf3-49d3-8bff-601cfc96c4a5","Type":"ContainerDied","Data":"3b241f4f0d61953fd8db9f03162f41c15f4640b20ed666996d54bbdd4f451966"} Nov 25 22:11:20 crc kubenswrapper[4931]: I1125 22:11:20.286951 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance311e-account-delete-fq67b" event={"ID":"31eac098-adf3-49d3-8bff-601cfc96c4a5","Type":"ContainerStarted","Data":"4e8c2915135f993ab80beac5f9bd28a6f2ec618c4099b4e966d1fdfcec269f55"} Nov 25 22:11:20 crc kubenswrapper[4931]: I1125 22:11:20.289673 4931 generic.go:334] "Generic (PLEG): container finished" podID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerID="64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee" exitCode=143 Nov 25 22:11:20 crc kubenswrapper[4931]: I1125 22:11:20.289732 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"338c52e4-b31c-48df-8a37-a1a36c9fadf9","Type":"ContainerDied","Data":"64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee"} Nov 25 22:11:20 crc kubenswrapper[4931]: I1125 22:11:20.430528 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa092118-ff34-40e3-8f23-bcc5dc2ae1d8" path="/var/lib/kubelet/pods/fa092118-ff34-40e3-8f23-bcc5dc2ae1d8/volumes" Nov 25 22:11:21 crc kubenswrapper[4931]: I1125 22:11:21.626505 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:21 crc kubenswrapper[4931]: I1125 22:11:21.745019 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eac098-adf3-49d3-8bff-601cfc96c4a5-operator-scripts\") pod \"31eac098-adf3-49d3-8bff-601cfc96c4a5\" (UID: \"31eac098-adf3-49d3-8bff-601cfc96c4a5\") " Nov 25 22:11:21 crc kubenswrapper[4931]: I1125 22:11:21.745190 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcslh\" (UniqueName: \"kubernetes.io/projected/31eac098-adf3-49d3-8bff-601cfc96c4a5-kube-api-access-kcslh\") pod \"31eac098-adf3-49d3-8bff-601cfc96c4a5\" (UID: \"31eac098-adf3-49d3-8bff-601cfc96c4a5\") " Nov 25 22:11:21 crc kubenswrapper[4931]: I1125 22:11:21.746508 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31eac098-adf3-49d3-8bff-601cfc96c4a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "31eac098-adf3-49d3-8bff-601cfc96c4a5" (UID: "31eac098-adf3-49d3-8bff-601cfc96c4a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:11:21 crc kubenswrapper[4931]: I1125 22:11:21.750589 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31eac098-adf3-49d3-8bff-601cfc96c4a5-kube-api-access-kcslh" (OuterVolumeSpecName: "kube-api-access-kcslh") pod "31eac098-adf3-49d3-8bff-601cfc96c4a5" (UID: "31eac098-adf3-49d3-8bff-601cfc96c4a5"). InnerVolumeSpecName "kube-api-access-kcslh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:11:21 crc kubenswrapper[4931]: I1125 22:11:21.846601 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcslh\" (UniqueName: \"kubernetes.io/projected/31eac098-adf3-49d3-8bff-601cfc96c4a5-kube-api-access-kcslh\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:21 crc kubenswrapper[4931]: I1125 22:11:21.846633 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31eac098-adf3-49d3-8bff-601cfc96c4a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:22 crc kubenswrapper[4931]: I1125 22:11:22.308463 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance311e-account-delete-fq67b" event={"ID":"31eac098-adf3-49d3-8bff-601cfc96c4a5","Type":"ContainerDied","Data":"4e8c2915135f993ab80beac5f9bd28a6f2ec618c4099b4e966d1fdfcec269f55"} Nov 25 22:11:22 crc kubenswrapper[4931]: I1125 22:11:22.308506 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e8c2915135f993ab80beac5f9bd28a6f2ec618c4099b4e966d1fdfcec269f55" Nov 25 22:11:22 crc kubenswrapper[4931]: I1125 22:11:22.308558 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance311e-account-delete-fq67b" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.129822 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.268491 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-internal-tls-certs\") pod \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.268549 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-logs\") pod \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.268582 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-public-tls-certs\") pod \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.268644 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-scripts\") pod \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.268702 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.268751 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-combined-ca-bundle\") pod \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.268778 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-config-data\") pod \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.268842 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgvkc\" (UniqueName: \"kubernetes.io/projected/338c52e4-b31c-48df-8a37-a1a36c9fadf9-kube-api-access-jgvkc\") pod \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.268889 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-httpd-run\") pod \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\" (UID: \"338c52e4-b31c-48df-8a37-a1a36c9fadf9\") " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.269586 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "338c52e4-b31c-48df-8a37-a1a36c9fadf9" (UID: "338c52e4-b31c-48df-8a37-a1a36c9fadf9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.270141 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-logs" (OuterVolumeSpecName: "logs") pod "338c52e4-b31c-48df-8a37-a1a36c9fadf9" (UID: "338c52e4-b31c-48df-8a37-a1a36c9fadf9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.275573 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-scripts" (OuterVolumeSpecName: "scripts") pod "338c52e4-b31c-48df-8a37-a1a36c9fadf9" (UID: "338c52e4-b31c-48df-8a37-a1a36c9fadf9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.276559 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/338c52e4-b31c-48df-8a37-a1a36c9fadf9-kube-api-access-jgvkc" (OuterVolumeSpecName: "kube-api-access-jgvkc") pod "338c52e4-b31c-48df-8a37-a1a36c9fadf9" (UID: "338c52e4-b31c-48df-8a37-a1a36c9fadf9"). InnerVolumeSpecName "kube-api-access-jgvkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.279031 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "338c52e4-b31c-48df-8a37-a1a36c9fadf9" (UID: "338c52e4-b31c-48df-8a37-a1a36c9fadf9"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.310745 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "338c52e4-b31c-48df-8a37-a1a36c9fadf9" (UID: "338c52e4-b31c-48df-8a37-a1a36c9fadf9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.314135 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "338c52e4-b31c-48df-8a37-a1a36c9fadf9" (UID: "338c52e4-b31c-48df-8a37-a1a36c9fadf9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.316482 4931 generic.go:334] "Generic (PLEG): container finished" podID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerID="4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be" exitCode=0 Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.316519 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"338c52e4-b31c-48df-8a37-a1a36c9fadf9","Type":"ContainerDied","Data":"4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be"} Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.316562 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"338c52e4-b31c-48df-8a37-a1a36c9fadf9","Type":"ContainerDied","Data":"06c8d3fe04995a69d7556d1ea41b0d356fb01e45e5b10aa40e9b8af8f4ed07be"} Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.316580 4931 scope.go:117] "RemoveContainer" containerID="4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.316583 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.320945 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "338c52e4-b31c-48df-8a37-a1a36c9fadf9" (UID: "338c52e4-b31c-48df-8a37-a1a36c9fadf9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.335704 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-config-data" (OuterVolumeSpecName: "config-data") pod "338c52e4-b31c-48df-8a37-a1a36c9fadf9" (UID: "338c52e4-b31c-48df-8a37-a1a36c9fadf9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.361771 4931 scope.go:117] "RemoveContainer" containerID="64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.370204 4931 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.370232 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.370265 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.370275 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.370284 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.370293 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgvkc\" (UniqueName: \"kubernetes.io/projected/338c52e4-b31c-48df-8a37-a1a36c9fadf9-kube-api-access-jgvkc\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.370304 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.370312 4931 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/338c52e4-b31c-48df-8a37-a1a36c9fadf9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.370320 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/338c52e4-b31c-48df-8a37-a1a36c9fadf9-logs\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.377735 4931 scope.go:117] "RemoveContainer" containerID="4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be" Nov 25 22:11:23 crc kubenswrapper[4931]: E1125 22:11:23.378416 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be\": container with ID starting with 4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be not found: ID does not exist" containerID="4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.378459 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be"} err="failed to get container status \"4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be\": rpc error: code = NotFound desc = could not find container \"4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be\": container with ID starting with 4b0270dc3a863cb84fe3c60d0c495f0ad34fb5d5e96b9bd3b9edf357ad76e0be not found: ID does not exist" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.378533 4931 scope.go:117] "RemoveContainer" containerID="64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee" Nov 25 22:11:23 crc kubenswrapper[4931]: E1125 22:11:23.378958 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee\": container with ID starting with 64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee not found: ID does not exist" containerID="64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.378983 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee"} err="failed to get container status \"64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee\": rpc error: code = NotFound desc = could not find container \"64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee\": container with ID starting with 64306723876c14e49bc9509b96dc52cadf1547474010f80a730169d301d277ee not found: ID does not exist" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.382426 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.472488 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.667511 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:11:23 crc kubenswrapper[4931]: I1125 22:11:23.676584 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.164439 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance311e-account-delete-fq67b"] Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.174179 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-311e-account-create-update-f5k5q"] Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.184397 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-x8gms"] Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.191853 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-311e-account-create-update-f5k5q"] Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.198750 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance311e-account-delete-fq67b"] Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.204538 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-x8gms"] Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.435027 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31eac098-adf3-49d3-8bff-601cfc96c4a5" path="/var/lib/kubelet/pods/31eac098-adf3-49d3-8bff-601cfc96c4a5/volumes" Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.436424 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" path="/var/lib/kubelet/pods/338c52e4-b31c-48df-8a37-a1a36c9fadf9/volumes" Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.437774 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84edc32e-99b0-4d11-8aff-b445d6a03786" path="/var/lib/kubelet/pods/84edc32e-99b0-4d11-8aff-b445d6a03786/volumes" Nov 25 22:11:24 crc kubenswrapper[4931]: I1125 22:11:24.439647 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6228ec7-9ffd-446c-b51c-99414b57969d" path="/var/lib/kubelet/pods/d6228ec7-9ffd-446c-b51c-99414b57969d/volumes" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.390381 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-8j7l7"] Nov 25 22:11:25 crc kubenswrapper[4931]: E1125 22:11:25.391163 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerName="glance-httpd" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.391187 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerName="glance-httpd" Nov 25 22:11:25 crc kubenswrapper[4931]: E1125 22:11:25.391219 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerName="glance-log" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.391231 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerName="glance-log" Nov 25 22:11:25 crc kubenswrapper[4931]: E1125 22:11:25.391248 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31eac098-adf3-49d3-8bff-601cfc96c4a5" containerName="mariadb-account-delete" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.391259 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="31eac098-adf3-49d3-8bff-601cfc96c4a5" containerName="mariadb-account-delete" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.391475 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerName="glance-httpd" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.391496 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="31eac098-adf3-49d3-8bff-601cfc96c4a5" containerName="mariadb-account-delete" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.391512 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="338c52e4-b31c-48df-8a37-a1a36c9fadf9" containerName="glance-log" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.392201 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.398710 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt"] Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.399954 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.402869 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.422632 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt"] Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.434506 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-8j7l7"] Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.500995 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ndkf\" (UniqueName: \"kubernetes.io/projected/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-kube-api-access-4ndkf\") pod \"glance-db-create-8j7l7\" (UID: \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\") " pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.501124 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa62d794-8196-47d9-bf6b-515ea4d1a35b-operator-scripts\") pod \"glance-6a8c-account-create-update-t2mvt\" (UID: \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\") " pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.501156 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-operator-scripts\") pod \"glance-db-create-8j7l7\" (UID: \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\") " pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.501218 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8hnb\" (UniqueName: \"kubernetes.io/projected/aa62d794-8196-47d9-bf6b-515ea4d1a35b-kube-api-access-p8hnb\") pod \"glance-6a8c-account-create-update-t2mvt\" (UID: \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\") " pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.603281 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8hnb\" (UniqueName: \"kubernetes.io/projected/aa62d794-8196-47d9-bf6b-515ea4d1a35b-kube-api-access-p8hnb\") pod \"glance-6a8c-account-create-update-t2mvt\" (UID: \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\") " pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.603416 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ndkf\" (UniqueName: \"kubernetes.io/projected/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-kube-api-access-4ndkf\") pod \"glance-db-create-8j7l7\" (UID: \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\") " pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.603479 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa62d794-8196-47d9-bf6b-515ea4d1a35b-operator-scripts\") pod \"glance-6a8c-account-create-update-t2mvt\" (UID: \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\") " pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.603519 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-operator-scripts\") pod \"glance-db-create-8j7l7\" (UID: \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\") " pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.605145 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-operator-scripts\") pod \"glance-db-create-8j7l7\" (UID: \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\") " pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.605525 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa62d794-8196-47d9-bf6b-515ea4d1a35b-operator-scripts\") pod \"glance-6a8c-account-create-update-t2mvt\" (UID: \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\") " pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.634986 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ndkf\" (UniqueName: \"kubernetes.io/projected/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-kube-api-access-4ndkf\") pod \"glance-db-create-8j7l7\" (UID: \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\") " pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.636529 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8hnb\" (UniqueName: \"kubernetes.io/projected/aa62d794-8196-47d9-bf6b-515ea4d1a35b-kube-api-access-p8hnb\") pod \"glance-6a8c-account-create-update-t2mvt\" (UID: \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\") " pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.715519 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:25 crc kubenswrapper[4931]: I1125 22:11:25.729597 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:26 crc kubenswrapper[4931]: I1125 22:11:26.008152 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-8j7l7"] Nov 25 22:11:26 crc kubenswrapper[4931]: W1125 22:11:26.012302 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3836796_a01c_4f07_a2e9_7dfc2b7054e1.slice/crio-129abfd8d97c13e03c35f053b60906e6e39d5675339ca628967814c6f825ce5f WatchSource:0}: Error finding container 129abfd8d97c13e03c35f053b60906e6e39d5675339ca628967814c6f825ce5f: Status 404 returned error can't find the container with id 129abfd8d97c13e03c35f053b60906e6e39d5675339ca628967814c6f825ce5f Nov 25 22:11:26 crc kubenswrapper[4931]: I1125 22:11:26.161007 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt"] Nov 25 22:11:26 crc kubenswrapper[4931]: W1125 22:11:26.163848 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa62d794_8196_47d9_bf6b_515ea4d1a35b.slice/crio-c9bd03500757321a668bd5e8359ca051311402cdc9a6463d2c55d13a4efb78d5 WatchSource:0}: Error finding container c9bd03500757321a668bd5e8359ca051311402cdc9a6463d2c55d13a4efb78d5: Status 404 returned error can't find the container with id c9bd03500757321a668bd5e8359ca051311402cdc9a6463d2c55d13a4efb78d5 Nov 25 22:11:26 crc kubenswrapper[4931]: I1125 22:11:26.344074 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" event={"ID":"aa62d794-8196-47d9-bf6b-515ea4d1a35b","Type":"ContainerStarted","Data":"902741845ac8789170899ec25fcba35ce5e4095b888512b806c7823cbca65d73"} Nov 25 22:11:26 crc kubenswrapper[4931]: I1125 22:11:26.344148 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" event={"ID":"aa62d794-8196-47d9-bf6b-515ea4d1a35b","Type":"ContainerStarted","Data":"c9bd03500757321a668bd5e8359ca051311402cdc9a6463d2c55d13a4efb78d5"} Nov 25 22:11:26 crc kubenswrapper[4931]: I1125 22:11:26.345912 4931 generic.go:334] "Generic (PLEG): container finished" podID="a3836796-a01c-4f07-a2e9-7dfc2b7054e1" containerID="bbe4cbf4f8bd07253e3590f255203ff3eb6ca5585c0a97e5c353f16a7abc5ea4" exitCode=0 Nov 25 22:11:26 crc kubenswrapper[4931]: I1125 22:11:26.345954 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-8j7l7" event={"ID":"a3836796-a01c-4f07-a2e9-7dfc2b7054e1","Type":"ContainerDied","Data":"bbe4cbf4f8bd07253e3590f255203ff3eb6ca5585c0a97e5c353f16a7abc5ea4"} Nov 25 22:11:26 crc kubenswrapper[4931]: I1125 22:11:26.345980 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-8j7l7" event={"ID":"a3836796-a01c-4f07-a2e9-7dfc2b7054e1","Type":"ContainerStarted","Data":"129abfd8d97c13e03c35f053b60906e6e39d5675339ca628967814c6f825ce5f"} Nov 25 22:11:26 crc kubenswrapper[4931]: I1125 22:11:26.366303 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" podStartSLOduration=1.366282598 podStartE2EDuration="1.366282598s" podCreationTimestamp="2025-11-25 22:11:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:11:26.360282305 +0000 UTC m=+1224.716794424" watchObservedRunningTime="2025-11-25 22:11:26.366282598 +0000 UTC m=+1224.722794727" Nov 25 22:11:27 crc kubenswrapper[4931]: I1125 22:11:27.358961 4931 generic.go:334] "Generic (PLEG): container finished" podID="aa62d794-8196-47d9-bf6b-515ea4d1a35b" containerID="902741845ac8789170899ec25fcba35ce5e4095b888512b806c7823cbca65d73" exitCode=0 Nov 25 22:11:27 crc kubenswrapper[4931]: I1125 22:11:27.359008 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" event={"ID":"aa62d794-8196-47d9-bf6b-515ea4d1a35b","Type":"ContainerDied","Data":"902741845ac8789170899ec25fcba35ce5e4095b888512b806c7823cbca65d73"} Nov 25 22:11:27 crc kubenswrapper[4931]: I1125 22:11:27.722058 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:27 crc kubenswrapper[4931]: I1125 22:11:27.836302 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ndkf\" (UniqueName: \"kubernetes.io/projected/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-kube-api-access-4ndkf\") pod \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\" (UID: \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\") " Nov 25 22:11:27 crc kubenswrapper[4931]: I1125 22:11:27.836417 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-operator-scripts\") pod \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\" (UID: \"a3836796-a01c-4f07-a2e9-7dfc2b7054e1\") " Nov 25 22:11:27 crc kubenswrapper[4931]: I1125 22:11:27.837430 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3836796-a01c-4f07-a2e9-7dfc2b7054e1" (UID: "a3836796-a01c-4f07-a2e9-7dfc2b7054e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:11:27 crc kubenswrapper[4931]: I1125 22:11:27.845389 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-kube-api-access-4ndkf" (OuterVolumeSpecName: "kube-api-access-4ndkf") pod "a3836796-a01c-4f07-a2e9-7dfc2b7054e1" (UID: "a3836796-a01c-4f07-a2e9-7dfc2b7054e1"). InnerVolumeSpecName "kube-api-access-4ndkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:11:27 crc kubenswrapper[4931]: I1125 22:11:27.937784 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ndkf\" (UniqueName: \"kubernetes.io/projected/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-kube-api-access-4ndkf\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:27 crc kubenswrapper[4931]: I1125 22:11:27.937820 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3836796-a01c-4f07-a2e9-7dfc2b7054e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.372317 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-8j7l7" event={"ID":"a3836796-a01c-4f07-a2e9-7dfc2b7054e1","Type":"ContainerDied","Data":"129abfd8d97c13e03c35f053b60906e6e39d5675339ca628967814c6f825ce5f"} Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.373014 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="129abfd8d97c13e03c35f053b60906e6e39d5675339ca628967814c6f825ce5f" Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.372464 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-8j7l7" Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.715007 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.750757 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa62d794-8196-47d9-bf6b-515ea4d1a35b-operator-scripts\") pod \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\" (UID: \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\") " Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.750848 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8hnb\" (UniqueName: \"kubernetes.io/projected/aa62d794-8196-47d9-bf6b-515ea4d1a35b-kube-api-access-p8hnb\") pod \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\" (UID: \"aa62d794-8196-47d9-bf6b-515ea4d1a35b\") " Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.752642 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa62d794-8196-47d9-bf6b-515ea4d1a35b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aa62d794-8196-47d9-bf6b-515ea4d1a35b" (UID: "aa62d794-8196-47d9-bf6b-515ea4d1a35b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.756671 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa62d794-8196-47d9-bf6b-515ea4d1a35b-kube-api-access-p8hnb" (OuterVolumeSpecName: "kube-api-access-p8hnb") pod "aa62d794-8196-47d9-bf6b-515ea4d1a35b" (UID: "aa62d794-8196-47d9-bf6b-515ea4d1a35b"). InnerVolumeSpecName "kube-api-access-p8hnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.852280 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa62d794-8196-47d9-bf6b-515ea4d1a35b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:28 crc kubenswrapper[4931]: I1125 22:11:28.852319 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8hnb\" (UniqueName: \"kubernetes.io/projected/aa62d794-8196-47d9-bf6b-515ea4d1a35b-kube-api-access-p8hnb\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:29 crc kubenswrapper[4931]: I1125 22:11:29.386713 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" event={"ID":"aa62d794-8196-47d9-bf6b-515ea4d1a35b","Type":"ContainerDied","Data":"c9bd03500757321a668bd5e8359ca051311402cdc9a6463d2c55d13a4efb78d5"} Nov 25 22:11:29 crc kubenswrapper[4931]: I1125 22:11:29.387074 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9bd03500757321a668bd5e8359ca051311402cdc9a6463d2c55d13a4efb78d5" Nov 25 22:11:29 crc kubenswrapper[4931]: I1125 22:11:29.386851 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.608652 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-5h92t"] Nov 25 22:11:30 crc kubenswrapper[4931]: E1125 22:11:30.609068 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3836796-a01c-4f07-a2e9-7dfc2b7054e1" containerName="mariadb-database-create" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.609096 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3836796-a01c-4f07-a2e9-7dfc2b7054e1" containerName="mariadb-database-create" Nov 25 22:11:30 crc kubenswrapper[4931]: E1125 22:11:30.609135 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa62d794-8196-47d9-bf6b-515ea4d1a35b" containerName="mariadb-account-create-update" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.609142 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa62d794-8196-47d9-bf6b-515ea4d1a35b" containerName="mariadb-account-create-update" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.609351 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3836796-a01c-4f07-a2e9-7dfc2b7054e1" containerName="mariadb-database-create" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.609365 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa62d794-8196-47d9-bf6b-515ea4d1a35b" containerName="mariadb-account-create-update" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.609987 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.612988 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.615682 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-gk9qz" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.634445 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5h92t"] Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.678219 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-db-sync-config-data\") pod \"glance-db-sync-5h92t\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.678280 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-config-data\") pod \"glance-db-sync-5h92t\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.678669 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv542\" (UniqueName: \"kubernetes.io/projected/5d7d25ea-0573-484d-bd94-e9f1527edf1f-kube-api-access-vv542\") pod \"glance-db-sync-5h92t\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.782944 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-db-sync-config-data\") pod \"glance-db-sync-5h92t\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.783407 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-config-data\") pod \"glance-db-sync-5h92t\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.783536 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv542\" (UniqueName: \"kubernetes.io/projected/5d7d25ea-0573-484d-bd94-e9f1527edf1f-kube-api-access-vv542\") pod \"glance-db-sync-5h92t\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.788503 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-db-sync-config-data\") pod \"glance-db-sync-5h92t\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.788982 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-config-data\") pod \"glance-db-sync-5h92t\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.806423 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv542\" (UniqueName: \"kubernetes.io/projected/5d7d25ea-0573-484d-bd94-e9f1527edf1f-kube-api-access-vv542\") pod \"glance-db-sync-5h92t\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:30 crc kubenswrapper[4931]: I1125 22:11:30.933782 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:31 crc kubenswrapper[4931]: I1125 22:11:31.370010 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5h92t"] Nov 25 22:11:31 crc kubenswrapper[4931]: W1125 22:11:31.379141 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d7d25ea_0573_484d_bd94_e9f1527edf1f.slice/crio-182215cbf434d7e3fcca446f633c34e1a6f554ec93edf5d9c0112e83d6a77d12 WatchSource:0}: Error finding container 182215cbf434d7e3fcca446f633c34e1a6f554ec93edf5d9c0112e83d6a77d12: Status 404 returned error can't find the container with id 182215cbf434d7e3fcca446f633c34e1a6f554ec93edf5d9c0112e83d6a77d12 Nov 25 22:11:31 crc kubenswrapper[4931]: I1125 22:11:31.412902 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5h92t" event={"ID":"5d7d25ea-0573-484d-bd94-e9f1527edf1f","Type":"ContainerStarted","Data":"182215cbf434d7e3fcca446f633c34e1a6f554ec93edf5d9c0112e83d6a77d12"} Nov 25 22:11:32 crc kubenswrapper[4931]: I1125 22:11:32.447889 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5h92t" event={"ID":"5d7d25ea-0573-484d-bd94-e9f1527edf1f","Type":"ContainerStarted","Data":"bb98a279a2d470505b2d878082f2849be02eb50a03d8bfeaaa8637c0540fa30c"} Nov 25 22:11:32 crc kubenswrapper[4931]: I1125 22:11:32.495901 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-5h92t" podStartSLOduration=2.495840006 podStartE2EDuration="2.495840006s" podCreationTimestamp="2025-11-25 22:11:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:11:32.487094299 +0000 UTC m=+1230.843606418" watchObservedRunningTime="2025-11-25 22:11:32.495840006 +0000 UTC m=+1230.852352155" Nov 25 22:11:35 crc kubenswrapper[4931]: I1125 22:11:35.473717 4931 generic.go:334] "Generic (PLEG): container finished" podID="5d7d25ea-0573-484d-bd94-e9f1527edf1f" containerID="bb98a279a2d470505b2d878082f2849be02eb50a03d8bfeaaa8637c0540fa30c" exitCode=0 Nov 25 22:11:35 crc kubenswrapper[4931]: I1125 22:11:35.473843 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5h92t" event={"ID":"5d7d25ea-0573-484d-bd94-e9f1527edf1f","Type":"ContainerDied","Data":"bb98a279a2d470505b2d878082f2849be02eb50a03d8bfeaaa8637c0540fa30c"} Nov 25 22:11:36 crc kubenswrapper[4931]: I1125 22:11:36.896354 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:36 crc kubenswrapper[4931]: I1125 22:11:36.976894 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-db-sync-config-data\") pod \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " Nov 25 22:11:36 crc kubenswrapper[4931]: I1125 22:11:36.977014 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-config-data\") pod \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " Nov 25 22:11:36 crc kubenswrapper[4931]: I1125 22:11:36.977152 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv542\" (UniqueName: \"kubernetes.io/projected/5d7d25ea-0573-484d-bd94-e9f1527edf1f-kube-api-access-vv542\") pod \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\" (UID: \"5d7d25ea-0573-484d-bd94-e9f1527edf1f\") " Nov 25 22:11:36 crc kubenswrapper[4931]: I1125 22:11:36.987517 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d7d25ea-0573-484d-bd94-e9f1527edf1f-kube-api-access-vv542" (OuterVolumeSpecName: "kube-api-access-vv542") pod "5d7d25ea-0573-484d-bd94-e9f1527edf1f" (UID: "5d7d25ea-0573-484d-bd94-e9f1527edf1f"). InnerVolumeSpecName "kube-api-access-vv542". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:11:36 crc kubenswrapper[4931]: I1125 22:11:36.989175 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5d7d25ea-0573-484d-bd94-e9f1527edf1f" (UID: "5d7d25ea-0573-484d-bd94-e9f1527edf1f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:37 crc kubenswrapper[4931]: I1125 22:11:37.036089 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-config-data" (OuterVolumeSpecName: "config-data") pod "5d7d25ea-0573-484d-bd94-e9f1527edf1f" (UID: "5d7d25ea-0573-484d-bd94-e9f1527edf1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:37 crc kubenswrapper[4931]: I1125 22:11:37.079213 4931 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:37 crc kubenswrapper[4931]: I1125 22:11:37.079284 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7d25ea-0573-484d-bd94-e9f1527edf1f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:37 crc kubenswrapper[4931]: I1125 22:11:37.079310 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv542\" (UniqueName: \"kubernetes.io/projected/5d7d25ea-0573-484d-bd94-e9f1527edf1f-kube-api-access-vv542\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:37 crc kubenswrapper[4931]: I1125 22:11:37.507338 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5h92t" Nov 25 22:11:37 crc kubenswrapper[4931]: I1125 22:11:37.506681 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5h92t" event={"ID":"5d7d25ea-0573-484d-bd94-e9f1527edf1f","Type":"ContainerDied","Data":"182215cbf434d7e3fcca446f633c34e1a6f554ec93edf5d9c0112e83d6a77d12"} Nov 25 22:11:37 crc kubenswrapper[4931]: I1125 22:11:37.507449 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="182215cbf434d7e3fcca446f633c34e1a6f554ec93edf5d9c0112e83d6a77d12" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.646899 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 22:11:38 crc kubenswrapper[4931]: E1125 22:11:38.647482 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7d25ea-0573-484d-bd94-e9f1527edf1f" containerName="glance-db-sync" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.647496 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7d25ea-0573-484d-bd94-e9f1527edf1f" containerName="glance-db-sync" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.647669 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7d25ea-0573-484d-bd94-e9f1527edf1f" containerName="glance-db-sync" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.648804 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.650691 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.651214 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-gk9qz" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.651327 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.669411 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.800836 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l8bx\" (UniqueName: \"kubernetes.io/projected/0987f7d6-37f0-48ea-8873-13cfa65fde9c-kube-api-access-8l8bx\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.800894 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0987f7d6-37f0-48ea-8873-13cfa65fde9c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.800919 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0987f7d6-37f0-48ea-8873-13cfa65fde9c-config-data\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.800942 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801098 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-dev\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801150 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-sys\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801186 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0987f7d6-37f0-48ea-8873-13cfa65fde9c-logs\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801276 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801323 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801433 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801511 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801546 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801565 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-run\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.801586 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0987f7d6-37f0-48ea-8873-13cfa65fde9c-scripts\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.843597 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.843657 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903006 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l8bx\" (UniqueName: \"kubernetes.io/projected/0987f7d6-37f0-48ea-8873-13cfa65fde9c-kube-api-access-8l8bx\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903074 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0987f7d6-37f0-48ea-8873-13cfa65fde9c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903145 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0987f7d6-37f0-48ea-8873-13cfa65fde9c-config-data\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903179 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903226 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-dev\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903250 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-sys\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903277 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0987f7d6-37f0-48ea-8873-13cfa65fde9c-logs\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903327 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903360 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903434 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-sys\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903498 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-dev\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903507 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903547 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903585 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903672 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0987f7d6-37f0-48ea-8873-13cfa65fde9c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903755 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903815 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903843 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903862 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-run\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.903914 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0987f7d6-37f0-48ea-8873-13cfa65fde9c-scripts\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.904089 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0987f7d6-37f0-48ea-8873-13cfa65fde9c-logs\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.904108 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.904100 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.904152 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0987f7d6-37f0-48ea-8873-13cfa65fde9c-run\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.904124 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.912417 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0987f7d6-37f0-48ea-8873-13cfa65fde9c-scripts\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.918249 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0987f7d6-37f0-48ea-8873-13cfa65fde9c-config-data\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.918681 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.921337 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l8bx\" (UniqueName: \"kubernetes.io/projected/0987f7d6-37f0-48ea-8873-13cfa65fde9c-kube-api-access-8l8bx\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.922236 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.924349 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.939481 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.941798 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 22:11:38 crc kubenswrapper[4931]: I1125 22:11:38.972257 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0987f7d6-37f0-48ea-8873-13cfa65fde9c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107125 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-logs\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107185 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107245 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107273 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-run\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107323 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107346 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107409 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107433 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krmml\" (UniqueName: \"kubernetes.io/projected/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-kube-api-access-krmml\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107473 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-dev\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107496 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107542 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107574 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-sys\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107669 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.107810 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209286 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-logs\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209584 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209703 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209757 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-run\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209780 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-logs\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209850 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209886 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-run\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209905 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209937 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209964 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.209972 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krmml\" (UniqueName: \"kubernetes.io/projected/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-kube-api-access-krmml\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210019 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-dev\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210052 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210056 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210146 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210231 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210480 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210509 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-sys\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210533 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210550 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210634 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210661 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-sys\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210713 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.210767 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-dev\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.211058 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.215246 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.228744 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.234331 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.243411 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krmml\" (UniqueName: \"kubernetes.io/projected/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-kube-api-access-krmml\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.247851 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.272144 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:39 crc kubenswrapper[4931]: I1125 22:11:39.324381 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:40 crc kubenswrapper[4931]: I1125 22:11:39.453295 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 22:11:40 crc kubenswrapper[4931]: I1125 22:11:39.536693 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 22:11:40 crc kubenswrapper[4931]: I1125 22:11:40.535577 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"0987f7d6-37f0-48ea-8873-13cfa65fde9c","Type":"ContainerStarted","Data":"5e39722bb839fa3d463ce86488dcaa5dbd66529ee747cc4b083a0a558eb07ff8"} Nov 25 22:11:40 crc kubenswrapper[4931]: I1125 22:11:40.536221 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"0987f7d6-37f0-48ea-8873-13cfa65fde9c","Type":"ContainerStarted","Data":"190a32dfe69051f76b8e455525c31c4f8dae705526858b3d4f64f77c02864c1a"} Nov 25 22:11:40 crc kubenswrapper[4931]: I1125 22:11:40.536236 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"0987f7d6-37f0-48ea-8873-13cfa65fde9c","Type":"ContainerStarted","Data":"59d2be13468e8a296110715ed319a666f8cead4a0fe249e6034f7ce4cbcd77fc"} Nov 25 22:11:40 crc kubenswrapper[4931]: I1125 22:11:40.536248 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"0987f7d6-37f0-48ea-8873-13cfa65fde9c","Type":"ContainerStarted","Data":"fc58388c3045ebd46d904b8034a5a84a80f2a65e8eacffe6837864ab6a6d85b2"} Nov 25 22:11:40 crc kubenswrapper[4931]: I1125 22:11:40.611616 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.611589289 podStartE2EDuration="2.611589289s" podCreationTimestamp="2025-11-25 22:11:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:11:40.589966807 +0000 UTC m=+1238.946478926" watchObservedRunningTime="2025-11-25 22:11:40.611589289 +0000 UTC m=+1238.968101408" Nov 25 22:11:40 crc kubenswrapper[4931]: I1125 22:11:40.615469 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 22:11:41 crc kubenswrapper[4931]: I1125 22:11:41.546553 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e","Type":"ContainerStarted","Data":"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259"} Nov 25 22:11:41 crc kubenswrapper[4931]: I1125 22:11:41.547137 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e","Type":"ContainerStarted","Data":"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b"} Nov 25 22:11:41 crc kubenswrapper[4931]: I1125 22:11:41.547177 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e","Type":"ContainerStarted","Data":"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38"} Nov 25 22:11:41 crc kubenswrapper[4931]: I1125 22:11:41.547129 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-api" containerID="cri-o://54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259" gracePeriod=30 Nov 25 22:11:41 crc kubenswrapper[4931]: I1125 22:11:41.547191 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-httpd" containerID="cri-o://4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b" gracePeriod=30 Nov 25 22:11:41 crc kubenswrapper[4931]: I1125 22:11:41.547195 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e","Type":"ContainerStarted","Data":"3fe1856035d20ca3e2d7e0d584cef019d55805758fe983c365dbaeda3e04af25"} Nov 25 22:11:41 crc kubenswrapper[4931]: I1125 22:11:41.547246 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-log" containerID="cri-o://3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38" gracePeriod=30 Nov 25 22:11:41 crc kubenswrapper[4931]: I1125 22:11:41.577983 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=4.577942017 podStartE2EDuration="4.577942017s" podCreationTimestamp="2025-11-25 22:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:11:41.576028926 +0000 UTC m=+1239.932541095" watchObservedRunningTime="2025-11-25 22:11:41.577942017 +0000 UTC m=+1239.934454146" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.055971 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.158844 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159190 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-nvme\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159214 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-var-locks-brick\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159230 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-sys\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159265 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-dev\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159303 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-run\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159330 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krmml\" (UniqueName: \"kubernetes.io/projected/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-kube-api-access-krmml\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159352 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-lib-modules\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159373 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-config-data\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159405 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-logs\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159423 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-httpd-run\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159453 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-iscsi\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159493 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.159513 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-scripts\") pod \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\" (UID: \"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e\") " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.160968 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.161001 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.161045 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.161097 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.161150 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-sys" (OuterVolumeSpecName: "sys") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.161193 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-dev" (OuterVolumeSpecName: "dev") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.161239 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-run" (OuterVolumeSpecName: "run") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.161554 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.161679 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-logs" (OuterVolumeSpecName: "logs") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.166664 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-scripts" (OuterVolumeSpecName: "scripts") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.166757 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-kube-api-access-krmml" (OuterVolumeSpecName: "kube-api-access-krmml") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "kube-api-access-krmml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.167516 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.179738 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261683 4931 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-dev\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261735 4931 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-run\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261755 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krmml\" (UniqueName: \"kubernetes.io/projected/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-kube-api-access-krmml\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261774 4931 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261792 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261808 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261826 4931 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261903 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261922 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261947 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261964 4931 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261980 4931 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.261998 4931 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-sys\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.266082 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-config-data" (OuterVolumeSpecName: "config-data") pod "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" (UID: "e83cbd0f-782c-4fb1-9c26-ed8e90011f0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.283654 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.286936 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.363795 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.363824 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.363833 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.558430 4931 generic.go:334] "Generic (PLEG): container finished" podID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerID="54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259" exitCode=143 Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.558624 4931 generic.go:334] "Generic (PLEG): container finished" podID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerID="4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b" exitCode=0 Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.558639 4931 generic.go:334] "Generic (PLEG): container finished" podID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerID="3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38" exitCode=143 Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.558566 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e","Type":"ContainerDied","Data":"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259"} Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.558684 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e","Type":"ContainerDied","Data":"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b"} Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.558678 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.558695 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e","Type":"ContainerDied","Data":"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38"} Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.558794 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"e83cbd0f-782c-4fb1-9c26-ed8e90011f0e","Type":"ContainerDied","Data":"3fe1856035d20ca3e2d7e0d584cef019d55805758fe983c365dbaeda3e04af25"} Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.558704 4931 scope.go:117] "RemoveContainer" containerID="54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.596586 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.598242 4931 scope.go:117] "RemoveContainer" containerID="4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.617589 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.623720 4931 scope.go:117] "RemoveContainer" containerID="3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.641686 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 22:11:42 crc kubenswrapper[4931]: E1125 22:11:42.642016 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-httpd" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.642039 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-httpd" Nov 25 22:11:42 crc kubenswrapper[4931]: E1125 22:11:42.642063 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-log" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.642072 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-log" Nov 25 22:11:42 crc kubenswrapper[4931]: E1125 22:11:42.642085 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-api" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.642093 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-api" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.642271 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-log" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.642290 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-httpd" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.642300 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" containerName="glance-api" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.643607 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.656761 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.658725 4931 scope.go:117] "RemoveContainer" containerID="54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259" Nov 25 22:11:42 crc kubenswrapper[4931]: E1125 22:11:42.692206 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259\": container with ID starting with 54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259 not found: ID does not exist" containerID="54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.692448 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259"} err="failed to get container status \"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259\": rpc error: code = NotFound desc = could not find container \"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259\": container with ID starting with 54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259 not found: ID does not exist" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.692591 4931 scope.go:117] "RemoveContainer" containerID="4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.692375 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 22:11:42 crc kubenswrapper[4931]: E1125 22:11:42.693452 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b\": container with ID starting with 4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b not found: ID does not exist" containerID="4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.693501 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b"} err="failed to get container status \"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b\": rpc error: code = NotFound desc = could not find container \"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b\": container with ID starting with 4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b not found: ID does not exist" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.693528 4931 scope.go:117] "RemoveContainer" containerID="3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38" Nov 25 22:11:42 crc kubenswrapper[4931]: E1125 22:11:42.693867 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38\": container with ID starting with 3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38 not found: ID does not exist" containerID="3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.693924 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38"} err="failed to get container status \"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38\": rpc error: code = NotFound desc = could not find container \"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38\": container with ID starting with 3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38 not found: ID does not exist" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.693940 4931 scope.go:117] "RemoveContainer" containerID="54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.694282 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259"} err="failed to get container status \"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259\": rpc error: code = NotFound desc = could not find container \"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259\": container with ID starting with 54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259 not found: ID does not exist" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.694304 4931 scope.go:117] "RemoveContainer" containerID="4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.694624 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b"} err="failed to get container status \"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b\": rpc error: code = NotFound desc = could not find container \"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b\": container with ID starting with 4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b not found: ID does not exist" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.694731 4931 scope.go:117] "RemoveContainer" containerID="3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.695927 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38"} err="failed to get container status \"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38\": rpc error: code = NotFound desc = could not find container \"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38\": container with ID starting with 3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38 not found: ID does not exist" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.696033 4931 scope.go:117] "RemoveContainer" containerID="54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.696372 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259"} err="failed to get container status \"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259\": rpc error: code = NotFound desc = could not find container \"54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259\": container with ID starting with 54172cc2ffea66f922b46b6d00f511bebfd8c45ac4a36ed84821b31af2f3b259 not found: ID does not exist" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.696456 4931 scope.go:117] "RemoveContainer" containerID="4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.696822 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b"} err="failed to get container status \"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b\": rpc error: code = NotFound desc = could not find container \"4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b\": container with ID starting with 4d702c2ee57fcfea8960a8bbb30bf83ccf08814cb7f8f2e06c6bff7da703d05b not found: ID does not exist" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.697015 4931 scope.go:117] "RemoveContainer" containerID="3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.697461 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38"} err="failed to get container status \"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38\": rpc error: code = NotFound desc = could not find container \"3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38\": container with ID starting with 3277d1825d9ddaa8e275f99c3df1fc9c312cded03c056390545822420ba48c38 not found: ID does not exist" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.777271 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.777490 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.777604 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.777688 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.777803 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.777926 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-run\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.778011 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cgcl\" (UniqueName: \"kubernetes.io/projected/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-kube-api-access-2cgcl\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.778074 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.778163 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.778239 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-sys\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.778297 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.778357 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.778420 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-dev\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.778500 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.879954 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.880172 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.880286 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.880389 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.880589 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.880689 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-run\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.880783 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cgcl\" (UniqueName: \"kubernetes.io/projected/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-kube-api-access-2cgcl\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.880864 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.880972 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.881099 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-sys\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.881203 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.881305 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.881410 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-dev\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.881536 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.881811 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.886998 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.887168 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.887992 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-sys\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.888244 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.888276 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.888350 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.888396 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-dev\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.888410 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.888429 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.888739 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-run\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.894839 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.900578 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.904170 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cgcl\" (UniqueName: \"kubernetes.io/projected/b7c2ca26-5323-4f31-ba89-a529dd5ddde0-kube-api-access-2cgcl\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.906743 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:42 crc kubenswrapper[4931]: I1125 22:11:42.907089 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7c2ca26-5323-4f31-ba89-a529dd5ddde0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:43 crc kubenswrapper[4931]: I1125 22:11:43.002649 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:43 crc kubenswrapper[4931]: I1125 22:11:43.230244 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 22:11:43 crc kubenswrapper[4931]: I1125 22:11:43.569226 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7c2ca26-5323-4f31-ba89-a529dd5ddde0","Type":"ContainerStarted","Data":"af4fd4af99dcffa7d14855ddbae20be6785c2f52e4e2117f879f3e5d4895fabf"} Nov 25 22:11:43 crc kubenswrapper[4931]: I1125 22:11:43.569575 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7c2ca26-5323-4f31-ba89-a529dd5ddde0","Type":"ContainerStarted","Data":"2f92134124e5b2c992142a0e8a5f9da14e92671ffd710450d2d7f53371969d17"} Nov 25 22:11:44 crc kubenswrapper[4931]: I1125 22:11:44.438078 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e83cbd0f-782c-4fb1-9c26-ed8e90011f0e" path="/var/lib/kubelet/pods/e83cbd0f-782c-4fb1-9c26-ed8e90011f0e/volumes" Nov 25 22:11:44 crc kubenswrapper[4931]: I1125 22:11:44.587617 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7c2ca26-5323-4f31-ba89-a529dd5ddde0","Type":"ContainerStarted","Data":"130997af2c0ab71aa166a1322daf82770293bad6b23525006823ce7f3d666c3e"} Nov 25 22:11:44 crc kubenswrapper[4931]: I1125 22:11:44.587677 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b7c2ca26-5323-4f31-ba89-a529dd5ddde0","Type":"ContainerStarted","Data":"b6e8e84064cd02a338943382db4ddcf9aef0eb8ac31edd8039c7f562fab74fc9"} Nov 25 22:11:44 crc kubenswrapper[4931]: I1125 22:11:44.624008 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.6239723169999998 podStartE2EDuration="2.623972317s" podCreationTimestamp="2025-11-25 22:11:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:11:44.618286474 +0000 UTC m=+1242.974798653" watchObservedRunningTime="2025-11-25 22:11:44.623972317 +0000 UTC m=+1242.980484516" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.272563 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.273319 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.273341 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.316041 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.321153 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.346323 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.656305 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.656368 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.656389 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.677262 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.679839 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:49 crc kubenswrapper[4931]: I1125 22:11:49.682248 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.003704 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.004193 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.004217 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.047122 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.052140 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.077621 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.692131 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.692193 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.692209 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.707869 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.709248 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:53 crc kubenswrapper[4931]: I1125 22:11:53.722625 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.185443 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.188565 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.201253 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.204690 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.210565 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.222042 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321366 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-dev\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321475 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321527 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-sys\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321558 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-dev\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321591 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-sys\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321650 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-scripts\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321694 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321737 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321838 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28023b78-ba9e-4dbe-9033-59df28b3437c-logs\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321908 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28023b78-ba9e-4dbe-9033-59df28b3437c-config-data\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321958 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.321997 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322123 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-run\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322178 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322311 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322364 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28023b78-ba9e-4dbe-9033-59df28b3437c-scripts\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322397 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322442 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322492 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322546 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322585 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-logs\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322620 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322721 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322835 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28023b78-ba9e-4dbe-9033-59df28b3437c-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322934 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8t42\" (UniqueName: \"kubernetes.io/projected/28023b78-ba9e-4dbe-9033-59df28b3437c-kube-api-access-n8t42\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.322971 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-config-data\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.323017 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-run\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.323040 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n4wn\" (UniqueName: \"kubernetes.io/projected/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-kube-api-access-7n4wn\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424392 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28023b78-ba9e-4dbe-9033-59df28b3437c-scripts\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424438 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424467 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424493 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424517 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424538 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-logs\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424560 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424582 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424602 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28023b78-ba9e-4dbe-9033-59df28b3437c-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424623 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8t42\" (UniqueName: \"kubernetes.io/projected/28023b78-ba9e-4dbe-9033-59df28b3437c-kube-api-access-n8t42\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424645 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-config-data\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424669 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-run\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424688 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n4wn\" (UniqueName: \"kubernetes.io/projected/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-kube-api-access-7n4wn\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424719 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-dev\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424739 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424765 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-sys\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424782 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-dev\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424801 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-sys\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424834 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-scripts\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424860 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424908 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424935 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28023b78-ba9e-4dbe-9033-59df28b3437c-logs\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424956 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28023b78-ba9e-4dbe-9033-59df28b3437c-config-data\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.424976 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.425000 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.425024 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-run\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.425043 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.425071 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.425409 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.425802 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-dev\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.426533 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.426594 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-sys\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.426667 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-dev\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.426737 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-sys\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.427482 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.427596 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.427653 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.427719 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.427908 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.428206 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28023b78-ba9e-4dbe-9033-59df28b3437c-logs\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.428217 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.428257 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.428287 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.428628 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-logs\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.428695 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.428752 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/28023b78-ba9e-4dbe-9033-59df28b3437c-run\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.429295 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28023b78-ba9e-4dbe-9033-59df28b3437c-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.429697 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-run\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.429776 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.429879 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.448766 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-scripts\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.449013 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28023b78-ba9e-4dbe-9033-59df28b3437c-config-data\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.454971 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-config-data\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.475563 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28023b78-ba9e-4dbe-9033-59df28b3437c-scripts\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.483549 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n4wn\" (UniqueName: \"kubernetes.io/projected/e68e5a87-bc0c-4683-a9c2-aac189ca3c6d-kube-api-access-7n4wn\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.521196 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.537241 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.554202 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.570150 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8t42\" (UniqueName: \"kubernetes.io/projected/28023b78-ba9e-4dbe-9033-59df28b3437c-kube-api-access-n8t42\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.577171 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-2\" (UID: \"28023b78-ba9e-4dbe-9033-59df28b3437c\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.832243 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:11:56 crc kubenswrapper[4931]: I1125 22:11:56.849464 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:11:57 crc kubenswrapper[4931]: I1125 22:11:57.226288 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 22:11:57 crc kubenswrapper[4931]: W1125 22:11:57.229592 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28023b78_ba9e_4dbe_9033_59df28b3437c.slice/crio-f221919f2380789420c51aa05aa4d2b2b95c611c522080521e8fa83b0206f2bd WatchSource:0}: Error finding container f221919f2380789420c51aa05aa4d2b2b95c611c522080521e8fa83b0206f2bd: Status 404 returned error can't find the container with id f221919f2380789420c51aa05aa4d2b2b95c611c522080521e8fa83b0206f2bd Nov 25 22:11:57 crc kubenswrapper[4931]: I1125 22:11:57.366269 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 22:11:57 crc kubenswrapper[4931]: W1125 22:11:57.376376 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode68e5a87_bc0c_4683_a9c2_aac189ca3c6d.slice/crio-9a9bbcd62a529b7844f4428e731718d0f6f2ebed694080b877da140606835b74 WatchSource:0}: Error finding container 9a9bbcd62a529b7844f4428e731718d0f6f2ebed694080b877da140606835b74: Status 404 returned error can't find the container with id 9a9bbcd62a529b7844f4428e731718d0f6f2ebed694080b877da140606835b74 Nov 25 22:11:57 crc kubenswrapper[4931]: I1125 22:11:57.729359 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"28023b78-ba9e-4dbe-9033-59df28b3437c","Type":"ContainerStarted","Data":"96581d9eb33bd9ce9749b16d586c3e601688d23c9f51e41b8f5f0f4de7b91649"} Nov 25 22:11:57 crc kubenswrapper[4931]: I1125 22:11:57.729793 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"28023b78-ba9e-4dbe-9033-59df28b3437c","Type":"ContainerStarted","Data":"948859728a2c6760d94984f0e5341ed61b4cdb4bbab932179910ed04b5a7d1e8"} Nov 25 22:11:57 crc kubenswrapper[4931]: I1125 22:11:57.729812 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"28023b78-ba9e-4dbe-9033-59df28b3437c","Type":"ContainerStarted","Data":"f221919f2380789420c51aa05aa4d2b2b95c611c522080521e8fa83b0206f2bd"} Nov 25 22:11:57 crc kubenswrapper[4931]: I1125 22:11:57.731684 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d","Type":"ContainerStarted","Data":"587153f23ada559e89c7934849e707a8672e116f7e41f7436aacf4ed2b07f3d6"} Nov 25 22:11:57 crc kubenswrapper[4931]: I1125 22:11:57.731744 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d","Type":"ContainerStarted","Data":"9a9bbcd62a529b7844f4428e731718d0f6f2ebed694080b877da140606835b74"} Nov 25 22:11:58 crc kubenswrapper[4931]: I1125 22:11:58.743161 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"28023b78-ba9e-4dbe-9033-59df28b3437c","Type":"ContainerStarted","Data":"d2af5c39ccc70dfc6d6130e29df83590e931bd9c9d75d80ace5793d23297af8b"} Nov 25 22:11:58 crc kubenswrapper[4931]: I1125 22:11:58.747148 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d","Type":"ContainerStarted","Data":"f9044c2d536bde39f1ee45dd464e620fcd11693af5c5bd5e0aea9056680b81b5"} Nov 25 22:11:58 crc kubenswrapper[4931]: I1125 22:11:58.747182 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e68e5a87-bc0c-4683-a9c2-aac189ca3c6d","Type":"ContainerStarted","Data":"cb321166c5436816516f1132428f28b6ff6062a257e40c0355133fd71f21aea3"} Nov 25 22:11:58 crc kubenswrapper[4931]: I1125 22:11:58.779275 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.7792552710000002 podStartE2EDuration="3.779255271s" podCreationTimestamp="2025-11-25 22:11:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:11:58.775295886 +0000 UTC m=+1257.131808035" watchObservedRunningTime="2025-11-25 22:11:58.779255271 +0000 UTC m=+1257.135767400" Nov 25 22:11:58 crc kubenswrapper[4931]: I1125 22:11:58.811191 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.81116127 podStartE2EDuration="3.81116127s" podCreationTimestamp="2025-11-25 22:11:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:11:58.801029028 +0000 UTC m=+1257.157541187" watchObservedRunningTime="2025-11-25 22:11:58.81116127 +0000 UTC m=+1257.167673429" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.832987 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.835018 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.835052 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.850287 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.850367 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.850396 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.875856 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.884698 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.891740 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.896997 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.897524 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.903219 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.904234 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.905057 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.919460 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:06 crc kubenswrapper[4931]: I1125 22:12:06.927016 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:07 crc kubenswrapper[4931]: I1125 22:12:07.897092 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:07 crc kubenswrapper[4931]: I1125 22:12:07.897778 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:07 crc kubenswrapper[4931]: I1125 22:12:07.897849 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:07 crc kubenswrapper[4931]: I1125 22:12:07.897868 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:07 crc kubenswrapper[4931]: I1125 22:12:07.903197 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:07 crc kubenswrapper[4931]: I1125 22:12:07.906460 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:07 crc kubenswrapper[4931]: I1125 22:12:07.906965 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 22:12:07 crc kubenswrapper[4931]: I1125 22:12:07.908327 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 22:12:08 crc kubenswrapper[4931]: I1125 22:12:08.843555 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:12:08 crc kubenswrapper[4931]: I1125 22:12:08.843632 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:12:08 crc kubenswrapper[4931]: I1125 22:12:08.843686 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 22:12:08 crc kubenswrapper[4931]: I1125 22:12:08.844365 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ac1334287ef510421b58306fba1e2a55e64de3a45349db9fac90d16b42c433c"} pod="openshift-machine-config-operator/machine-config-daemon-djjhg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 22:12:08 crc kubenswrapper[4931]: I1125 22:12:08.844429 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" containerID="cri-o://9ac1334287ef510421b58306fba1e2a55e64de3a45349db9fac90d16b42c433c" gracePeriod=600 Nov 25 22:12:09 crc kubenswrapper[4931]: I1125 22:12:09.923477 4931 generic.go:334] "Generic (PLEG): container finished" podID="e22448de-5122-43be-853a-081347c13cf2" containerID="9ac1334287ef510421b58306fba1e2a55e64de3a45349db9fac90d16b42c433c" exitCode=0 Nov 25 22:12:09 crc kubenswrapper[4931]: I1125 22:12:09.923548 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerDied","Data":"9ac1334287ef510421b58306fba1e2a55e64de3a45349db9fac90d16b42c433c"} Nov 25 22:12:09 crc kubenswrapper[4931]: I1125 22:12:09.924008 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"90db9e7682e94b8d54bbdd5d154989aeb94726c14a1c13fff2838dc83db574a2"} Nov 25 22:12:09 crc kubenswrapper[4931]: I1125 22:12:09.924049 4931 scope.go:117] "RemoveContainer" containerID="f284a26c69ccc770f75121f3dbe719003137727be2b936d7ebbd69a2ec135257" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.631768 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.633869 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.641066 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.641609 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.644454 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.677780 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.677908 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.780238 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.780318 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.780430 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.821254 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:12 crc kubenswrapper[4931]: I1125 22:14:12.991455 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:13 crc kubenswrapper[4931]: I1125 22:14:13.524312 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 22:14:14 crc kubenswrapper[4931]: I1125 22:14:14.107542 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a4815387-1a2e-4d9d-9347-28c4935ef8e6","Type":"ContainerStarted","Data":"b4de2f960d9e36d69c78d7b3649dcc694b4f4e96d1d5ddcd6bee67fa652474ff"} Nov 25 22:14:14 crc kubenswrapper[4931]: I1125 22:14:14.107596 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a4815387-1a2e-4d9d-9347-28c4935ef8e6","Type":"ContainerStarted","Data":"30b6e033af177d2bafc99f713982ea0e566a2f85d3d3ec8d6b0f34a68c0a4889"} Nov 25 22:14:14 crc kubenswrapper[4931]: I1125 22:14:14.122567 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.122531639 podStartE2EDuration="2.122531639s" podCreationTimestamp="2025-11-25 22:14:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:14:14.121478021 +0000 UTC m=+1392.477990150" watchObservedRunningTime="2025-11-25 22:14:14.122531639 +0000 UTC m=+1392.479043758" Nov 25 22:14:15 crc kubenswrapper[4931]: I1125 22:14:15.131935 4931 generic.go:334] "Generic (PLEG): container finished" podID="a4815387-1a2e-4d9d-9347-28c4935ef8e6" containerID="b4de2f960d9e36d69c78d7b3649dcc694b4f4e96d1d5ddcd6bee67fa652474ff" exitCode=0 Nov 25 22:14:15 crc kubenswrapper[4931]: I1125 22:14:15.132078 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a4815387-1a2e-4d9d-9347-28c4935ef8e6","Type":"ContainerDied","Data":"b4de2f960d9e36d69c78d7b3649dcc694b4f4e96d1d5ddcd6bee67fa652474ff"} Nov 25 22:14:16 crc kubenswrapper[4931]: I1125 22:14:16.491989 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:16 crc kubenswrapper[4931]: I1125 22:14:16.552126 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kubelet-dir\") pod \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\" (UID: \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\") " Nov 25 22:14:16 crc kubenswrapper[4931]: I1125 22:14:16.552271 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kube-api-access\") pod \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\" (UID: \"a4815387-1a2e-4d9d-9347-28c4935ef8e6\") " Nov 25 22:14:16 crc kubenswrapper[4931]: I1125 22:14:16.552285 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a4815387-1a2e-4d9d-9347-28c4935ef8e6" (UID: "a4815387-1a2e-4d9d-9347-28c4935ef8e6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:14:16 crc kubenswrapper[4931]: I1125 22:14:16.552947 4931 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 22:14:16 crc kubenswrapper[4931]: I1125 22:14:16.562253 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a4815387-1a2e-4d9d-9347-28c4935ef8e6" (UID: "a4815387-1a2e-4d9d-9347-28c4935ef8e6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:14:16 crc kubenswrapper[4931]: I1125 22:14:16.655256 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4815387-1a2e-4d9d-9347-28c4935ef8e6-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.154351 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a4815387-1a2e-4d9d-9347-28c4935ef8e6","Type":"ContainerDied","Data":"30b6e033af177d2bafc99f713982ea0e566a2f85d3d3ec8d6b0f34a68c0a4889"} Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.154405 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30b6e033af177d2bafc99f713982ea0e566a2f85d3d3ec8d6b0f34a68c0a4889" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.154487 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.424445 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 22:14:17 crc kubenswrapper[4931]: E1125 22:14:17.424804 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4815387-1a2e-4d9d-9347-28c4935ef8e6" containerName="pruner" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.424820 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4815387-1a2e-4d9d-9347-28c4935ef8e6" containerName="pruner" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.425061 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4815387-1a2e-4d9d-9347-28c4935ef8e6" containerName="pruner" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.425621 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.428910 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.433586 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.443607 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.470899 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.470986 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kube-api-access\") pod \"installer-9-crc\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.471103 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-var-lock\") pod \"installer-9-crc\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.572556 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-var-lock\") pod \"installer-9-crc\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.572651 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.572657 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-var-lock\") pod \"installer-9-crc\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.572684 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kube-api-access\") pod \"installer-9-crc\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.572786 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.593440 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kube-api-access\") pod \"installer-9-crc\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:17 crc kubenswrapper[4931]: I1125 22:14:17.755306 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:18 crc kubenswrapper[4931]: I1125 22:14:18.247161 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 22:14:19 crc kubenswrapper[4931]: I1125 22:14:19.175126 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ce336cf2-d251-4e2c-8879-2e7b48eaed18","Type":"ContainerStarted","Data":"418690a0290211eae426103498e5ee7063bf4c21b25a041d3043dcbfbf748246"} Nov 25 22:14:19 crc kubenswrapper[4931]: I1125 22:14:19.175181 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ce336cf2-d251-4e2c-8879-2e7b48eaed18","Type":"ContainerStarted","Data":"2288e12c095d4b795196b13dcc5ba12e69e506c66bf90b2d9dba7f643488442d"} Nov 25 22:14:19 crc kubenswrapper[4931]: I1125 22:14:19.198732 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.198704965 podStartE2EDuration="2.198704965s" podCreationTimestamp="2025-11-25 22:14:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:14:19.194220464 +0000 UTC m=+1397.550732583" watchObservedRunningTime="2025-11-25 22:14:19.198704965 +0000 UTC m=+1397.555217124" Nov 25 22:14:38 crc kubenswrapper[4931]: I1125 22:14:38.843857 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:14:38 crc kubenswrapper[4931]: I1125 22:14:38.844281 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.587620 4931 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.589074 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.590083 4931 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.590321 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35" gracePeriod=15 Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.590365 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940" gracePeriod=15 Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.590450 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b" gracePeriod=15 Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.590494 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d" gracePeriod=15 Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.590527 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924" gracePeriod=15 Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.592544 4931 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 22:14:56 crc kubenswrapper[4931]: E1125 22:14:56.592943 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.592957 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 22:14:56 crc kubenswrapper[4931]: E1125 22:14:56.592977 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.592985 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 22:14:56 crc kubenswrapper[4931]: E1125 22:14:56.593016 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593025 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 22:14:56 crc kubenswrapper[4931]: E1125 22:14:56.593059 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593067 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 22:14:56 crc kubenswrapper[4931]: E1125 22:14:56.593088 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593098 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 22:14:56 crc kubenswrapper[4931]: E1125 22:14:56.593124 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593132 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593296 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593311 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593329 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593343 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593352 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593362 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 22:14:56 crc kubenswrapper[4931]: E1125 22:14:56.593534 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.593543 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.656047 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.715598 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.715655 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.715704 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.715720 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.715758 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.715776 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.715804 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.715833 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.817687 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.817783 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.817841 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.817895 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.817957 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818025 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818069 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818028 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818215 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818254 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818400 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818407 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818448 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818443 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818491 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.818514 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: I1125 22:14:56.940340 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:14:56 crc kubenswrapper[4931]: E1125 22:14:56.973478 4931 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b5fb1f52805bd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 22:14:56.972572093 +0000 UTC m=+1435.329084212,LastTimestamp:2025-11-25 22:14:56.972572093 +0000 UTC m=+1435.329084212,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.251476 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.252255 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.253029 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.253517 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.254137 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.254184 4931 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.254602 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="200ms" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.456538 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="400ms" Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.545379 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f"} Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.545437 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"67e550d59328be26dea3dda00d51994cc9415fe7d89edc2d3fc4330e0d1cdbce"} Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.546195 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.546787 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.547745 4931 generic.go:334] "Generic (PLEG): container finished" podID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" containerID="418690a0290211eae426103498e5ee7063bf4c21b25a041d3043dcbfbf748246" exitCode=0 Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.547830 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ce336cf2-d251-4e2c-8879-2e7b48eaed18","Type":"ContainerDied","Data":"418690a0290211eae426103498e5ee7063bf4c21b25a041d3043dcbfbf748246"} Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.548661 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.549222 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.549634 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.551156 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.552337 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.552957 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940" exitCode=0 Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.552982 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b" exitCode=0 Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.552994 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d" exitCode=0 Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.553005 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924" exitCode=2 Nov 25 22:14:57 crc kubenswrapper[4931]: I1125 22:14:57.553042 4931 scope.go:117] "RemoveContainer" containerID="faefab31b4d340ebb89cb01f15c41d0c257fb3d57f2058bca50c370f190656dd" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.733610 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:14:57Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:14:57Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:14:57Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T22:14:57Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.733865 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.734220 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.734569 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.735145 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.735249 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 22:14:57 crc kubenswrapper[4931]: E1125 22:14:57.858176 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="800ms" Nov 25 22:14:58 crc kubenswrapper[4931]: I1125 22:14:58.568615 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 22:14:58 crc kubenswrapper[4931]: E1125 22:14:58.659358 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="1.6s" Nov 25 22:14:58 crc kubenswrapper[4931]: I1125 22:14:58.940559 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:58 crc kubenswrapper[4931]: I1125 22:14:58.941806 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:58 crc kubenswrapper[4931]: I1125 22:14:58.942208 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.065539 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kube-api-access\") pod \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.065747 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kubelet-dir\") pod \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.065836 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ce336cf2-d251-4e2c-8879-2e7b48eaed18" (UID: "ce336cf2-d251-4e2c-8879-2e7b48eaed18"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.065899 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-var-lock" (OuterVolumeSpecName: "var-lock") pod "ce336cf2-d251-4e2c-8879-2e7b48eaed18" (UID: "ce336cf2-d251-4e2c-8879-2e7b48eaed18"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.065918 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-var-lock\") pod \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\" (UID: \"ce336cf2-d251-4e2c-8879-2e7b48eaed18\") " Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.066286 4931 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.066309 4931 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce336cf2-d251-4e2c-8879-2e7b48eaed18-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.074788 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ce336cf2-d251-4e2c-8879-2e7b48eaed18" (UID: "ce336cf2-d251-4e2c-8879-2e7b48eaed18"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.075548 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.076348 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.076948 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.077232 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.077558 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.167532 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.167583 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.167642 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.167748 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.167779 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.167837 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.168205 4931 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.168237 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce336cf2-d251-4e2c-8879-2e7b48eaed18-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.168259 4931 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.168277 4931 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.581654 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.583134 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35" exitCode=0 Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.583236 4931 scope.go:117] "RemoveContainer" containerID="15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.583325 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.585224 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ce336cf2-d251-4e2c-8879-2e7b48eaed18","Type":"ContainerDied","Data":"2288e12c095d4b795196b13dcc5ba12e69e506c66bf90b2d9dba7f643488442d"} Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.585259 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2288e12c095d4b795196b13dcc5ba12e69e506c66bf90b2d9dba7f643488442d" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.585290 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.611381 4931 scope.go:117] "RemoveContainer" containerID="2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.611906 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.612109 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.612747 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.613070 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.613860 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.614051 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.642195 4931 scope.go:117] "RemoveContainer" containerID="231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.667705 4931 scope.go:117] "RemoveContainer" containerID="b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.690612 4931 scope.go:117] "RemoveContainer" containerID="166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.730491 4931 scope.go:117] "RemoveContainer" containerID="f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.754930 4931 scope.go:117] "RemoveContainer" containerID="15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940" Nov 25 22:14:59 crc kubenswrapper[4931]: E1125 22:14:59.756011 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\": container with ID starting with 15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940 not found: ID does not exist" containerID="15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.756049 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940"} err="failed to get container status \"15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\": rpc error: code = NotFound desc = could not find container \"15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940\": container with ID starting with 15060e4770a2684c16acfc13f1c5e2a1884b322eda728b76988f1dd5d8fc3940 not found: ID does not exist" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.756084 4931 scope.go:117] "RemoveContainer" containerID="2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b" Nov 25 22:14:59 crc kubenswrapper[4931]: E1125 22:14:59.756850 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\": container with ID starting with 2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b not found: ID does not exist" containerID="2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.756944 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b"} err="failed to get container status \"2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\": rpc error: code = NotFound desc = could not find container \"2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b\": container with ID starting with 2d5660e233bee954058cfda6d3725255e3670ff5d7e66f8138083ad1ac57915b not found: ID does not exist" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.756966 4931 scope.go:117] "RemoveContainer" containerID="231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d" Nov 25 22:14:59 crc kubenswrapper[4931]: E1125 22:14:59.758313 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\": container with ID starting with 231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d not found: ID does not exist" containerID="231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.758347 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d"} err="failed to get container status \"231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\": rpc error: code = NotFound desc = could not find container \"231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d\": container with ID starting with 231373aa0faf240e9a7a653b528a9dec0f57bc2a5cb241293fa93933c0c4114d not found: ID does not exist" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.758361 4931 scope.go:117] "RemoveContainer" containerID="b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924" Nov 25 22:14:59 crc kubenswrapper[4931]: E1125 22:14:59.758624 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\": container with ID starting with b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924 not found: ID does not exist" containerID="b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.758654 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924"} err="failed to get container status \"b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\": rpc error: code = NotFound desc = could not find container \"b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924\": container with ID starting with b124c5d8fdad77645af45552e6baeeaea483eae3a147564ef85cdd66c126a924 not found: ID does not exist" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.758681 4931 scope.go:117] "RemoveContainer" containerID="166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35" Nov 25 22:14:59 crc kubenswrapper[4931]: E1125 22:14:59.759856 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\": container with ID starting with 166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35 not found: ID does not exist" containerID="166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.759897 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35"} err="failed to get container status \"166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\": rpc error: code = NotFound desc = could not find container \"166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35\": container with ID starting with 166483b408971089f0fff354bccffc6fc1a2904f1eb4b8463aefe91071805c35 not found: ID does not exist" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.759913 4931 scope.go:117] "RemoveContainer" containerID="f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c" Nov 25 22:14:59 crc kubenswrapper[4931]: E1125 22:14:59.760261 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\": container with ID starting with f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c not found: ID does not exist" containerID="f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c" Nov 25 22:14:59 crc kubenswrapper[4931]: I1125 22:14:59.760291 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c"} err="failed to get container status \"f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\": rpc error: code = NotFound desc = could not find container \"f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c\": container with ID starting with f1da10ccd5549cebda6b51b81c4f5a86151a8e136412cc8fcd990197519ec04c not found: ID does not exist" Nov 25 22:15:00 crc kubenswrapper[4931]: E1125 22:15:00.260602 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="3.2s" Nov 25 22:15:00 crc kubenswrapper[4931]: I1125 22:15:00.433584 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 22:15:00 crc kubenswrapper[4931]: E1125 22:15:00.716456 4931 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b5fb1f52805bd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 22:14:56.972572093 +0000 UTC m=+1435.329084212,LastTimestamp:2025-11-25 22:14:56.972572093 +0000 UTC m=+1435.329084212,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 22:15:02 crc kubenswrapper[4931]: I1125 22:15:02.437133 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:02 crc kubenswrapper[4931]: I1125 22:15:02.437553 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:03 crc kubenswrapper[4931]: E1125 22:15:03.462322 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="6.4s" Nov 25 22:15:06 crc kubenswrapper[4931]: E1125 22:15:06.439629 4931 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC glance-kuttl-tests/mysql-db-openstack-galera-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/persistentvolumeclaims/mysql-db-openstack-galera-0\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="glance-kuttl-tests/openstack-galera-0" volumeName="mysql-db" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.422596 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.423997 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.424565 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: E1125 22:15:07.458866 4931 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC glance-kuttl-tests/mysql-db-openstack-galera-1: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/persistentvolumeclaims/mysql-db-openstack-galera-1\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="glance-kuttl-tests/openstack-galera-1" volumeName="mysql-db" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.463016 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.463075 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:07 crc kubenswrapper[4931]: E1125 22:15:07.463669 4931 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.464718 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.645292 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.16:8080/readyz\": dial tcp 10.217.0.16:8080: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.676258 4931 generic.go:334] "Generic (PLEG): container finished" podID="952695c7-9eff-4e33-8777-6a5ea50a2b94" containerID="e124225a0871c7f1bdd8ce73d78764c990f06ae1ed18feb2546f4f0cfbd20973" exitCode=1 Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.676367 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" event={"ID":"952695c7-9eff-4e33-8777-6a5ea50a2b94","Type":"ContainerDied","Data":"e124225a0871c7f1bdd8ce73d78764c990f06ae1ed18feb2546f4f0cfbd20973"} Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.677727 4931 scope.go:117] "RemoveContainer" containerID="e124225a0871c7f1bdd8ce73d78764c990f06ae1ed18feb2546f4f0cfbd20973" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.679960 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.680316 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.680916 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.683697 4931 generic.go:334] "Generic (PLEG): container finished" podID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" containerID="a00ec071c6e2988ead68259cb195dfb80083380991f1ec8272588984bd389df1" exitCode=1 Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.683768 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" event={"ID":"60182b4c-52b4-47a9-8c5a-ab0274136ef1","Type":"ContainerDied","Data":"a00ec071c6e2988ead68259cb195dfb80083380991f1ec8272588984bd389df1"} Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.684274 4931 scope.go:117] "RemoveContainer" containerID="a00ec071c6e2988ead68259cb195dfb80083380991f1ec8272588984bd389df1" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.685102 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.685402 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.685732 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.686143 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.688983 4931 generic.go:334] "Generic (PLEG): container finished" podID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" containerID="4908a63e7b813b51872992e0b6f626b33e2adac29589994682430011e21c8be5" exitCode=1 Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.689089 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" event={"ID":"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef","Type":"ContainerDied","Data":"4908a63e7b813b51872992e0b6f626b33e2adac29589994682430011e21c8be5"} Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.689935 4931 scope.go:117] "RemoveContainer" containerID="4908a63e7b813b51872992e0b6f626b33e2adac29589994682430011e21c8be5" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.690017 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.691152 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.691668 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.692133 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.692600 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.693362 4931 generic.go:334] "Generic (PLEG): container finished" podID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" containerID="6f212b00aeed7232535f9c5ae6ef70d445c379d0b5f19f7dd4fdbe485d5236b1" exitCode=1 Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.693450 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" event={"ID":"7b1bb711-ce82-4ea4-9667-ff4816b162f4","Type":"ContainerDied","Data":"6f212b00aeed7232535f9c5ae6ef70d445c379d0b5f19f7dd4fdbe485d5236b1"} Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.693965 4931 scope.go:117] "RemoveContainer" containerID="6f212b00aeed7232535f9c5ae6ef70d445c379d0b5f19f7dd4fdbe485d5236b1" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.694147 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.694569 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.695963 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.696063 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a8eaff8c130c084c8bde76f2ecf49155eb2d7028794836783dcb2a7c0964cafd"} Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.696303 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.696744 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.697093 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.698724 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4575978-a78e-42b2-a2dc-1052e5777b04" containerID="49f7a9f920f39948ab4b6feb74172179b41a03f604bebb7cb1d51e2f0adb0762" exitCode=1 Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.698754 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" event={"ID":"f4575978-a78e-42b2-a2dc-1052e5777b04","Type":"ContainerDied","Data":"49f7a9f920f39948ab4b6feb74172179b41a03f604bebb7cb1d51e2f0adb0762"} Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.699195 4931 scope.go:117] "RemoveContainer" containerID="49f7a9f920f39948ab4b6feb74172179b41a03f604bebb7cb1d51e2f0adb0762" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.699870 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.700482 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6545c6c9d6-7rkzq\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.700946 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.701347 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.701898 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.702217 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:07 crc kubenswrapper[4931]: I1125 22:15:07.702485 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.711801 4931 generic.go:334] "Generic (PLEG): container finished" podID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" containerID="09f97af7d5cc06c9c4a41f29fef5e7f36a1967b505c8993f82a24a93d1fe74d7" exitCode=1 Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.711919 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" event={"ID":"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef","Type":"ContainerDied","Data":"09f97af7d5cc06c9c4a41f29fef5e7f36a1967b505c8993f82a24a93d1fe74d7"} Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.712190 4931 scope.go:117] "RemoveContainer" containerID="4908a63e7b813b51872992e0b6f626b33e2adac29589994682430011e21c8be5" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.712727 4931 scope.go:117] "RemoveContainer" containerID="09f97af7d5cc06c9c4a41f29fef5e7f36a1967b505c8993f82a24a93d1fe74d7" Nov 25 22:15:08 crc kubenswrapper[4931]: E1125 22:15:08.713052 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-6fb5bb54f-chfmg_openstack-operators(0700c2d1-0a31-4d46-9fe7-4bdb809d1cef)\"" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.713249 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.714219 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.715318 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6545c6c9d6-7rkzq\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.716076 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.716156 4931 generic.go:334] "Generic (PLEG): container finished" podID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" containerID="6ec289a5d9baaead242659a3a1ddd3bdd720fcf3c3949038853a9500b101aca1" exitCode=1 Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.716213 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" event={"ID":"7b1bb711-ce82-4ea4-9667-ff4816b162f4","Type":"ContainerDied","Data":"6ec289a5d9baaead242659a3a1ddd3bdd720fcf3c3949038853a9500b101aca1"} Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.716583 4931 scope.go:117] "RemoveContainer" containerID="6ec289a5d9baaead242659a3a1ddd3bdd720fcf3c3949038853a9500b101aca1" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.716613 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: E1125 22:15:08.716802 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-zsd98_openstack-operators(7b1bb711-ce82-4ea4-9667-ff4816b162f4)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.717593 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.718953 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.720032 4931 generic.go:334] "Generic (PLEG): container finished" podID="044c2639-0ceb-4be3-8326-4fdc552aefb6" containerID="a51a02fea4208b822ab73a8a98ebe332fbe70871d27bff531648f01d2a486f6f" exitCode=1 Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.720162 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" event={"ID":"044c2639-0ceb-4be3-8326-4fdc552aefb6","Type":"ContainerDied","Data":"a51a02fea4208b822ab73a8a98ebe332fbe70871d27bff531648f01d2a486f6f"} Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.721157 4931 scope.go:117] "RemoveContainer" containerID="a51a02fea4208b822ab73a8a98ebe332fbe70871d27bff531648f01d2a486f6f" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.719739 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.722485 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6545c6c9d6-7rkzq\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.722824 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.723159 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.723364 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.723611 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.723840 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.724263 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.724529 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6545c6c9d6-7rkzq\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.725060 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.725663 4931 status_manager.go:851] "Failed to get status for pod" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-5b59697cf5-jthhc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.726167 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.726593 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.726857 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.727158 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.728820 4931 generic.go:334] "Generic (PLEG): container finished" podID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" containerID="70c61a03853f7dba64993162bb7cd6128959f992ed3f36626f2bf1f5c54971e2" exitCode=1 Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.729029 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" event={"ID":"60182b4c-52b4-47a9-8c5a-ab0274136ef1","Type":"ContainerDied","Data":"70c61a03853f7dba64993162bb7cd6128959f992ed3f36626f2bf1f5c54971e2"} Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.729182 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.87:8081/readyz\": dial tcp 10.217.0.87:8081: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.729273 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.87:8081/healthz\": dial tcp 10.217.0.87:8081: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.729503 4931 scope.go:117] "RemoveContainer" containerID="70c61a03853f7dba64993162bb7cd6128959f992ed3f36626f2bf1f5c54971e2" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.729679 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: E1125 22:15:08.729730 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-5bfcf68b5b-79jjz_metallb-system(60182b4c-52b4-47a9-8c5a-ab0274136ef1)\"" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.730199 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.730465 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6545c6c9d6-7rkzq\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.730654 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.730899 4931 status_manager.go:851] "Failed to get status for pod" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-5b59697cf5-jthhc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.731101 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.731290 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.732247 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.736930 4931 generic.go:334] "Generic (PLEG): container finished" podID="8f67c56d-e216-435c-8384-91d28a4d864e" containerID="21a452c30a9e676b35b4edc64252a39b87cab9ce149dfd4d212b6915c586d372" exitCode=1 Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.737009 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" event={"ID":"8f67c56d-e216-435c-8384-91d28a4d864e","Type":"ContainerDied","Data":"21a452c30a9e676b35b4edc64252a39b87cab9ce149dfd4d212b6915c586d372"} Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.738395 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.738818 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6545c6c9d6-7rkzq\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.738885 4931 scope.go:117] "RemoveContainer" containerID="21a452c30a9e676b35b4edc64252a39b87cab9ce149dfd4d212b6915c586d372" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.739141 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.740663 4931 status_manager.go:851] "Failed to get status for pod" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-5b59697cf5-jthhc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.740911 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.741139 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.741353 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.741546 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.741746 4931 status_manager.go:851] "Failed to get status for pod" podUID="8f67c56d-e216-435c-8384-91d28a4d864e" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-7cb66c858-rgrsz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.755768 4931 generic.go:334] "Generic (PLEG): container finished" podID="952695c7-9eff-4e33-8777-6a5ea50a2b94" containerID="4adddbe541d62d39b690cbdd0b4d15800aeb0e40c4d2aeaa648225bd63e9909e" exitCode=1 Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.756216 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" event={"ID":"952695c7-9eff-4e33-8777-6a5ea50a2b94","Type":"ContainerDied","Data":"4adddbe541d62d39b690cbdd0b4d15800aeb0e40c4d2aeaa648225bd63e9909e"} Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.757325 4931 scope.go:117] "RemoveContainer" containerID="4adddbe541d62d39b690cbdd0b4d15800aeb0e40c4d2aeaa648225bd63e9909e" Nov 25 22:15:08 crc kubenswrapper[4931]: E1125 22:15:08.757747 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-7b69997d9b-sz4qv_openstack-operators(952695c7-9eff-4e33-8777-6a5ea50a2b94)\"" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.759009 4931 status_manager.go:851] "Failed to get status for pod" podUID="8f67c56d-e216-435c-8384-91d28a4d864e" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-7cb66c858-rgrsz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.759471 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.759823 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6545c6c9d6-7rkzq\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.760034 4931 generic.go:334] "Generic (PLEG): container finished" podID="4e0590b1-c897-4382-bc93-4a1b0653a526" containerID="af490f90de1d00d98315fa959270aa4c4c0a7da562d4c067edccd71ebd5b709a" exitCode=1 Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.760145 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.760163 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" event={"ID":"4e0590b1-c897-4382-bc93-4a1b0653a526","Type":"ContainerDied","Data":"af490f90de1d00d98315fa959270aa4c4c0a7da562d4c067edccd71ebd5b709a"} Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.761564 4931 scope.go:117] "RemoveContainer" containerID="af490f90de1d00d98315fa959270aa4c4c0a7da562d4c067edccd71ebd5b709a" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.762201 4931 status_manager.go:851] "Failed to get status for pod" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-5b59697cf5-jthhc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.762522 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.762847 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.763180 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.763622 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.764085 4931 status_manager.go:851] "Failed to get status for pod" podUID="8f67c56d-e216-435c-8384-91d28a4d864e" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-7cb66c858-rgrsz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.764224 4931 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="ce430f93deb52debf13f6b594b2cde8a11092be492c715305d88cf405a09d697" exitCode=0 Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.764369 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"ce430f93deb52debf13f6b594b2cde8a11092be492c715305d88cf405a09d697"} Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.764534 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.764605 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.764627 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:08 crc kubenswrapper[4931]: E1125 22:15:08.764922 4931 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.765126 4931 status_manager.go:851] "Failed to get status for pod" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-6bfcf864bd-cfmpv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.766177 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6545c6c9d6-7rkzq\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.766552 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.766871 4931 status_manager.go:851] "Failed to get status for pod" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-5b59697cf5-jthhc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.767177 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.767366 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4575978-a78e-42b2-a2dc-1052e5777b04" containerID="114c22ad6168b9801bb71949516dd301af63efa6f115db00991a17a18f85d537" exitCode=1 Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.767405 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" event={"ID":"f4575978-a78e-42b2-a2dc-1052e5777b04","Type":"ContainerDied","Data":"114c22ad6168b9801bb71949516dd301af63efa6f115db00991a17a18f85d537"} Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.767461 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.767740 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.767795 4931 scope.go:117] "RemoveContainer" containerID="114c22ad6168b9801bb71949516dd301af63efa6f115db00991a17a18f85d537" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.768062 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: E1125 22:15:08.768100 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-6545c6c9d6-7rkzq_openstack-operators(f4575978-a78e-42b2-a2dc-1052e5777b04)\"" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.768252 4931 scope.go:117] "RemoveContainer" containerID="6f212b00aeed7232535f9c5ae6ef70d445c379d0b5f19f7dd4fdbe485d5236b1" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.768585 4931 status_manager.go:851] "Failed to get status for pod" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-6bfcf864bd-cfmpv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.768835 4931 status_manager.go:851] "Failed to get status for pod" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6fb5bb54f-chfmg\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.769465 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6545c6c9d6-7rkzq\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.779116 4931 status_manager.go:851] "Failed to get status for pod" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.780185 4931 status_manager.go:851] "Failed to get status for pod" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-5b59697cf5-jthhc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.780660 4931 status_manager.go:851] "Failed to get status for pod" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-5bfcf68b5b-79jjz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.781218 4931 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.781688 4931 status_manager.go:851] "Failed to get status for pod" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7b69997d9b-sz4qv\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.782050 4931 status_manager.go:851] "Failed to get status for pod" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-zsd98\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.782588 4931 status_manager.go:851] "Failed to get status for pod" podUID="8f67c56d-e216-435c-8384-91d28a4d864e" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-7cb66c858-rgrsz\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.843707 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.843759 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.891023 4931 scope.go:117] "RemoveContainer" containerID="a00ec071c6e2988ead68259cb195dfb80083380991f1ec8272588984bd389df1" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.919926 4931 scope.go:117] "RemoveContainer" containerID="e124225a0871c7f1bdd8ce73d78764c990f06ae1ed18feb2546f4f0cfbd20973" Nov 25 22:15:08 crc kubenswrapper[4931]: I1125 22:15:08.957762 4931 scope.go:117] "RemoveContainer" containerID="49f7a9f920f39948ab4b6feb74172179b41a03f604bebb7cb1d51e2f0adb0762" Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.782030 4931 generic.go:334] "Generic (PLEG): container finished" podID="044c2639-0ceb-4be3-8326-4fdc552aefb6" containerID="17f5fe9e4fcaa1ba0bce2ade9fbb0f823bb201cf3c31a62d5dea83030f4c865a" exitCode=1 Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.782141 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" event={"ID":"044c2639-0ceb-4be3-8326-4fdc552aefb6","Type":"ContainerDied","Data":"17f5fe9e4fcaa1ba0bce2ade9fbb0f823bb201cf3c31a62d5dea83030f4c865a"} Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.782588 4931 scope.go:117] "RemoveContainer" containerID="a51a02fea4208b822ab73a8a98ebe332fbe70871d27bff531648f01d2a486f6f" Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.783367 4931 scope.go:117] "RemoveContainer" containerID="17f5fe9e4fcaa1ba0bce2ade9fbb0f823bb201cf3c31a62d5dea83030f4c865a" Nov 25 22:15:09 crc kubenswrapper[4931]: E1125 22:15:09.783824 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-5b59697cf5-jthhc_openstack-operators(044c2639-0ceb-4be3-8326-4fdc552aefb6)\"" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.801362 4931 generic.go:334] "Generic (PLEG): container finished" podID="4e0590b1-c897-4382-bc93-4a1b0653a526" containerID="eeb63f50db0576037f45cd3a87c84ead941968a4dbade547d108e365512ad15c" exitCode=1 Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.801426 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" event={"ID":"4e0590b1-c897-4382-bc93-4a1b0653a526","Type":"ContainerDied","Data":"eeb63f50db0576037f45cd3a87c84ead941968a4dbade547d108e365512ad15c"} Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.801752 4931 scope.go:117] "RemoveContainer" containerID="eeb63f50db0576037f45cd3a87c84ead941968a4dbade547d108e365512ad15c" Nov 25 22:15:09 crc kubenswrapper[4931]: E1125 22:15:09.801960 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-6bfcf864bd-cfmpv_openstack-operators(4e0590b1-c897-4382-bc93-4a1b0653a526)\"" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.804710 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"933c2b1e835f1f7660590bf33cdfc12bd42522d36276872676cf703237ff1634"} Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.804788 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"312a3b7ac2b8d4573d541cc3bd351412ce1e8ae25d5ce0693b65a7f1a6b9e386"} Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.804829 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b1d94e9047547ad0df0ea76f8a851080112fec7f522cef5b78a9f873371eef36"} Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.819047 4931 generic.go:334] "Generic (PLEG): container finished" podID="8f67c56d-e216-435c-8384-91d28a4d864e" containerID="aad2a0ccd44a04dfc3d4fe5995c72793b181be8d649b2fdec3ab99d00c317efa" exitCode=1 Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.819105 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" event={"ID":"8f67c56d-e216-435c-8384-91d28a4d864e","Type":"ContainerDied","Data":"aad2a0ccd44a04dfc3d4fe5995c72793b181be8d649b2fdec3ab99d00c317efa"} Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.819676 4931 scope.go:117] "RemoveContainer" containerID="aad2a0ccd44a04dfc3d4fe5995c72793b181be8d649b2fdec3ab99d00c317efa" Nov 25 22:15:09 crc kubenswrapper[4931]: E1125 22:15:09.819868 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-7cb66c858-rgrsz_openstack-operators(8f67c56d-e216-435c-8384-91d28a4d864e)\"" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" podUID="8f67c56d-e216-435c-8384-91d28a4d864e" Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.829472 4931 scope.go:117] "RemoveContainer" containerID="af490f90de1d00d98315fa959270aa4c4c0a7da562d4c067edccd71ebd5b709a" Nov 25 22:15:09 crc kubenswrapper[4931]: I1125 22:15:09.888142 4931 scope.go:117] "RemoveContainer" containerID="21a452c30a9e676b35b4edc64252a39b87cab9ce149dfd4d212b6915c586d372" Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.475497 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.477289 4931 scope.go:117] "RemoveContainer" containerID="09f97af7d5cc06c9c4a41f29fef5e7f36a1967b505c8993f82a24a93d1fe74d7" Nov 25 22:15:10 crc kubenswrapper[4931]: E1125 22:15:10.477857 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-6fb5bb54f-chfmg_openstack-operators(0700c2d1-0a31-4d46-9fe7-4bdb809d1cef)\"" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.834239 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b4231810c5d3310a938cbc8b7bc96d2ebba8aa2764d6c18234f54b6f3d5e611e"} Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.834274 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f826e6388832376b03aa0badf5e06b83198fcd5163de2ab678ce981cf1735dd8"} Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.834480 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.834493 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.834669 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.840317 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.840362 4931 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d" exitCode=1 Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.840414 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d"} Nov 25 22:15:10 crc kubenswrapper[4931]: I1125 22:15:10.840987 4931 scope.go:117] "RemoveContainer" containerID="e2240c7719c0fad6b21f2770eafb444acc1731da86c1155f5e36060f0d254a1d" Nov 25 22:15:11 crc kubenswrapper[4931]: I1125 22:15:11.108497 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:15:11 crc kubenswrapper[4931]: I1125 22:15:11.855069 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 22:15:11 crc kubenswrapper[4931]: I1125 22:15:11.855330 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"358e093d45a0eb142d7a08833696c043b0f3af7b9f14acd4ab4895163340de4c"} Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.115468 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.115536 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.116406 4931 scope.go:117] "RemoveContainer" containerID="4adddbe541d62d39b690cbdd0b4d15800aeb0e40c4d2aeaa648225bd63e9909e" Nov 25 22:15:12 crc kubenswrapper[4931]: E1125 22:15:12.116755 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-7b69997d9b-sz4qv_openstack-operators(952695c7-9eff-4e33-8777-6a5ea50a2b94)\"" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.221284 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.227735 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.465087 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.465150 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.473584 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.497541 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.498312 4931 scope.go:117] "RemoveContainer" containerID="114c22ad6168b9801bb71949516dd301af63efa6f115db00991a17a18f85d537" Nov 25 22:15:12 crc kubenswrapper[4931]: E1125 22:15:12.498748 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-6545c6c9d6-7rkzq_openstack-operators(f4575978-a78e-42b2-a2dc-1052e5777b04)\"" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" Nov 25 22:15:12 crc kubenswrapper[4931]: I1125 22:15:12.864254 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:15:15 crc kubenswrapper[4931]: I1125 22:15:15.254380 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:15:15 crc kubenswrapper[4931]: I1125 22:15:15.254454 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:15:15 crc kubenswrapper[4931]: I1125 22:15:15.255059 4931 scope.go:117] "RemoveContainer" containerID="aad2a0ccd44a04dfc3d4fe5995c72793b181be8d649b2fdec3ab99d00c317efa" Nov 25 22:15:15 crc kubenswrapper[4931]: E1125 22:15:15.255414 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-7cb66c858-rgrsz_openstack-operators(8f67c56d-e216-435c-8384-91d28a4d864e)\"" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" podUID="8f67c56d-e216-435c-8384-91d28a4d864e" Nov 25 22:15:15 crc kubenswrapper[4931]: I1125 22:15:15.847497 4931 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:15 crc kubenswrapper[4931]: I1125 22:15:15.889900 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:15 crc kubenswrapper[4931]: I1125 22:15:15.889936 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:15 crc kubenswrapper[4931]: I1125 22:15:15.917092 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:16 crc kubenswrapper[4931]: I1125 22:15:16.059287 4931 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="897b709a-9491-4894-a2cc-f7b0154f8f73" Nov 25 22:15:16 crc kubenswrapper[4931]: I1125 22:15:16.900023 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:16 crc kubenswrapper[4931]: I1125 22:15:16.900331 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dee48871-9ee2-4083-a5be-427a390aaf20" Nov 25 22:15:16 crc kubenswrapper[4931]: I1125 22:15:16.903506 4931 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="897b709a-9491-4894-a2cc-f7b0154f8f73" Nov 25 22:15:17 crc kubenswrapper[4931]: I1125 22:15:17.303928 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:15:17 crc kubenswrapper[4931]: I1125 22:15:17.304512 4931 scope.go:117] "RemoveContainer" containerID="17f5fe9e4fcaa1ba0bce2ade9fbb0f823bb201cf3c31a62d5dea83030f4c865a" Nov 25 22:15:17 crc kubenswrapper[4931]: E1125 22:15:17.304702 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-5b59697cf5-jthhc_openstack-operators(044c2639-0ceb-4be3-8326-4fdc552aefb6)\"" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" Nov 25 22:15:17 crc kubenswrapper[4931]: I1125 22:15:17.645164 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:15:17 crc kubenswrapper[4931]: I1125 22:15:17.645969 4931 scope.go:117] "RemoveContainer" containerID="70c61a03853f7dba64993162bb7cd6128959f992ed3f36626f2bf1f5c54971e2" Nov 25 22:15:17 crc kubenswrapper[4931]: E1125 22:15:17.646234 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-5bfcf68b5b-79jjz_metallb-system(60182b4c-52b4-47a9-8c5a-ab0274136ef1)\"" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" Nov 25 22:15:18 crc kubenswrapper[4931]: I1125 22:15:18.728851 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:15:18 crc kubenswrapper[4931]: I1125 22:15:18.729438 4931 scope.go:117] "RemoveContainer" containerID="eeb63f50db0576037f45cd3a87c84ead941968a4dbade547d108e365512ad15c" Nov 25 22:15:18 crc kubenswrapper[4931]: E1125 22:15:18.729697 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-6bfcf864bd-cfmpv_openstack-operators(4e0590b1-c897-4382-bc93-4a1b0653a526)\"" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" Nov 25 22:15:20 crc kubenswrapper[4931]: I1125 22:15:20.423245 4931 scope.go:117] "RemoveContainer" containerID="6ec289a5d9baaead242659a3a1ddd3bdd720fcf3c3949038853a9500b101aca1" Nov 25 22:15:20 crc kubenswrapper[4931]: I1125 22:15:20.475934 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:15:20 crc kubenswrapper[4931]: I1125 22:15:20.476860 4931 scope.go:117] "RemoveContainer" containerID="09f97af7d5cc06c9c4a41f29fef5e7f36a1967b505c8993f82a24a93d1fe74d7" Nov 25 22:15:20 crc kubenswrapper[4931]: I1125 22:15:20.941609 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" event={"ID":"7b1bb711-ce82-4ea4-9667-ff4816b162f4","Type":"ContainerStarted","Data":"a4d5b21a4f10a37aa32a653d8f79b6eb0cabf19aaebf2e569325f4591a3cb8cd"} Nov 25 22:15:20 crc kubenswrapper[4931]: I1125 22:15:20.946010 4931 generic.go:334] "Generic (PLEG): container finished" podID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" containerID="f60a9d2362e3a3d4af899782755bda29a3f265b7103e5c9662a47be6301a5695" exitCode=1 Nov 25 22:15:20 crc kubenswrapper[4931]: I1125 22:15:20.946058 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" event={"ID":"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef","Type":"ContainerDied","Data":"f60a9d2362e3a3d4af899782755bda29a3f265b7103e5c9662a47be6301a5695"} Nov 25 22:15:20 crc kubenswrapper[4931]: I1125 22:15:20.946093 4931 scope.go:117] "RemoveContainer" containerID="09f97af7d5cc06c9c4a41f29fef5e7f36a1967b505c8993f82a24a93d1fe74d7" Nov 25 22:15:20 crc kubenswrapper[4931]: I1125 22:15:20.946779 4931 scope.go:117] "RemoveContainer" containerID="f60a9d2362e3a3d4af899782755bda29a3f265b7103e5c9662a47be6301a5695" Nov 25 22:15:20 crc kubenswrapper[4931]: E1125 22:15:20.947397 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-6fb5bb54f-chfmg_openstack-operators(0700c2d1-0a31-4d46-9fe7-4bdb809d1cef)\"" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" Nov 25 22:15:21 crc kubenswrapper[4931]: I1125 22:15:21.112850 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 22:15:21 crc kubenswrapper[4931]: I1125 22:15:21.966083 4931 generic.go:334] "Generic (PLEG): container finished" podID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" containerID="a4d5b21a4f10a37aa32a653d8f79b6eb0cabf19aaebf2e569325f4591a3cb8cd" exitCode=1 Nov 25 22:15:21 crc kubenswrapper[4931]: I1125 22:15:21.966159 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" event={"ID":"7b1bb711-ce82-4ea4-9667-ff4816b162f4","Type":"ContainerDied","Data":"a4d5b21a4f10a37aa32a653d8f79b6eb0cabf19aaebf2e569325f4591a3cb8cd"} Nov 25 22:15:21 crc kubenswrapper[4931]: I1125 22:15:21.966214 4931 scope.go:117] "RemoveContainer" containerID="6ec289a5d9baaead242659a3a1ddd3bdd720fcf3c3949038853a9500b101aca1" Nov 25 22:15:21 crc kubenswrapper[4931]: I1125 22:15:21.967588 4931 scope.go:117] "RemoveContainer" containerID="a4d5b21a4f10a37aa32a653d8f79b6eb0cabf19aaebf2e569325f4591a3cb8cd" Nov 25 22:15:21 crc kubenswrapper[4931]: E1125 22:15:21.968729 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-zsd98_openstack-operators(7b1bb711-ce82-4ea4-9667-ff4816b162f4)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" Nov 25 22:15:22 crc kubenswrapper[4931]: I1125 22:15:22.497637 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:15:22 crc kubenswrapper[4931]: I1125 22:15:22.498940 4931 scope.go:117] "RemoveContainer" containerID="114c22ad6168b9801bb71949516dd301af63efa6f115db00991a17a18f85d537" Nov 25 22:15:22 crc kubenswrapper[4931]: I1125 22:15:22.979616 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4575978-a78e-42b2-a2dc-1052e5777b04" containerID="28120008f68ab8447c744736d83346f31aae6356363b5efe208cea7a702c784d" exitCode=1 Nov 25 22:15:22 crc kubenswrapper[4931]: I1125 22:15:22.979732 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" event={"ID":"f4575978-a78e-42b2-a2dc-1052e5777b04","Type":"ContainerDied","Data":"28120008f68ab8447c744736d83346f31aae6356363b5efe208cea7a702c784d"} Nov 25 22:15:22 crc kubenswrapper[4931]: I1125 22:15:22.980209 4931 scope.go:117] "RemoveContainer" containerID="114c22ad6168b9801bb71949516dd301af63efa6f115db00991a17a18f85d537" Nov 25 22:15:22 crc kubenswrapper[4931]: I1125 22:15:22.981169 4931 scope.go:117] "RemoveContainer" containerID="28120008f68ab8447c744736d83346f31aae6356363b5efe208cea7a702c784d" Nov 25 22:15:22 crc kubenswrapper[4931]: E1125 22:15:22.981672 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-6545c6c9d6-7rkzq_openstack-operators(f4575978-a78e-42b2-a2dc-1052e5777b04)\"" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" Nov 25 22:15:24 crc kubenswrapper[4931]: I1125 22:15:24.388725 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 22:15:26 crc kubenswrapper[4931]: I1125 22:15:26.195427 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 22:15:26 crc kubenswrapper[4931]: I1125 22:15:26.423321 4931 scope.go:117] "RemoveContainer" containerID="4adddbe541d62d39b690cbdd0b4d15800aeb0e40c4d2aeaa648225bd63e9909e" Nov 25 22:15:26 crc kubenswrapper[4931]: I1125 22:15:26.458322 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 22:15:26 crc kubenswrapper[4931]: I1125 22:15:26.680171 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 22:15:26 crc kubenswrapper[4931]: I1125 22:15:26.763062 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.049202 4931 generic.go:334] "Generic (PLEG): container finished" podID="952695c7-9eff-4e33-8777-6a5ea50a2b94" containerID="31ac46a90802ccc16582ea23119e002f6ae960d59b6b9deb90d377b0fafc85f1" exitCode=1 Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.049271 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" event={"ID":"952695c7-9eff-4e33-8777-6a5ea50a2b94","Type":"ContainerDied","Data":"31ac46a90802ccc16582ea23119e002f6ae960d59b6b9deb90d377b0fafc85f1"} Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.049327 4931 scope.go:117] "RemoveContainer" containerID="4adddbe541d62d39b690cbdd0b4d15800aeb0e40c4d2aeaa648225bd63e9909e" Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.050183 4931 scope.go:117] "RemoveContainer" containerID="31ac46a90802ccc16582ea23119e002f6ae960d59b6b9deb90d377b0fafc85f1" Nov 25 22:15:27 crc kubenswrapper[4931]: E1125 22:15:27.050584 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-7b69997d9b-sz4qv_openstack-operators(952695c7-9eff-4e33-8777-6a5ea50a2b94)\"" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.303405 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.304257 4931 scope.go:117] "RemoveContainer" containerID="17f5fe9e4fcaa1ba0bce2ade9fbb0f823bb201cf3c31a62d5dea83030f4c865a" Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.375578 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.474008 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.534441 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.580422 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 22:15:27 crc kubenswrapper[4931]: I1125 22:15:27.731581 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.062697 4931 generic.go:334] "Generic (PLEG): container finished" podID="044c2639-0ceb-4be3-8326-4fdc552aefb6" containerID="7ba26f11563c6cc856c2d50ebc7125f560b7121a5d9f738fa872a9002aa33dda" exitCode=1 Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.062797 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" event={"ID":"044c2639-0ceb-4be3-8326-4fdc552aefb6","Type":"ContainerDied","Data":"7ba26f11563c6cc856c2d50ebc7125f560b7121a5d9f738fa872a9002aa33dda"} Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.063260 4931 scope.go:117] "RemoveContainer" containerID="17f5fe9e4fcaa1ba0bce2ade9fbb0f823bb201cf3c31a62d5dea83030f4c865a" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.065168 4931 scope.go:117] "RemoveContainer" containerID="7ba26f11563c6cc856c2d50ebc7125f560b7121a5d9f738fa872a9002aa33dda" Nov 25 22:15:28 crc kubenswrapper[4931]: E1125 22:15:28.065676 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-5b59697cf5-jthhc_openstack-operators(044c2639-0ceb-4be3-8326-4fdc552aefb6)\"" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.407766 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-44wj8" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.423325 4931 scope.go:117] "RemoveContainer" containerID="aad2a0ccd44a04dfc3d4fe5995c72793b181be8d649b2fdec3ab99d00c317efa" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.431226 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.575698 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.641185 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.688796 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.728533 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.729283 4931 scope.go:117] "RemoveContainer" containerID="eeb63f50db0576037f45cd3a87c84ead941968a4dbade547d108e365512ad15c" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.794381 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 22:15:28 crc kubenswrapper[4931]: I1125 22:15:28.939815 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.026993 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.090626 4931 generic.go:334] "Generic (PLEG): container finished" podID="8f67c56d-e216-435c-8384-91d28a4d864e" containerID="512092cb5bab7d49bb094278372ccd67706619dc39df920d362e11aec70170ba" exitCode=1 Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.090747 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" event={"ID":"8f67c56d-e216-435c-8384-91d28a4d864e","Type":"ContainerDied","Data":"512092cb5bab7d49bb094278372ccd67706619dc39df920d362e11aec70170ba"} Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.090825 4931 scope.go:117] "RemoveContainer" containerID="aad2a0ccd44a04dfc3d4fe5995c72793b181be8d649b2fdec3ab99d00c317efa" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.091449 4931 scope.go:117] "RemoveContainer" containerID="512092cb5bab7d49bb094278372ccd67706619dc39df920d362e11aec70170ba" Nov 25 22:15:29 crc kubenswrapper[4931]: E1125 22:15:29.091706 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-7cb66c858-rgrsz_openstack-operators(8f67c56d-e216-435c-8384-91d28a4d864e)\"" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" podUID="8f67c56d-e216-435c-8384-91d28a4d864e" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.100286 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" event={"ID":"4e0590b1-c897-4382-bc93-4a1b0653a526","Type":"ContainerStarted","Data":"da4b6146230b9b514eef62c76d77f4e26423619e78cf37d68290061de8e60376"} Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.100561 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.116541 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.328299 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.370583 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.391239 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.423612 4931 scope.go:117] "RemoveContainer" containerID="70c61a03853f7dba64993162bb7cd6128959f992ed3f36626f2bf1f5c54971e2" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.439000 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.458847 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-prgr5" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.503229 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.547211 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.573671 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.586566 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.588129 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.594405 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.868546 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.901849 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.905848 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 22:15:29 crc kubenswrapper[4931]: I1125 22:15:29.916277 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.024172 4931 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.115515 4931 generic.go:334] "Generic (PLEG): container finished" podID="4e0590b1-c897-4382-bc93-4a1b0653a526" containerID="da4b6146230b9b514eef62c76d77f4e26423619e78cf37d68290061de8e60376" exitCode=1 Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.115633 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" event={"ID":"4e0590b1-c897-4382-bc93-4a1b0653a526","Type":"ContainerDied","Data":"da4b6146230b9b514eef62c76d77f4e26423619e78cf37d68290061de8e60376"} Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.115735 4931 scope.go:117] "RemoveContainer" containerID="eeb63f50db0576037f45cd3a87c84ead941968a4dbade547d108e365512ad15c" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.116449 4931 scope.go:117] "RemoveContainer" containerID="da4b6146230b9b514eef62c76d77f4e26423619e78cf37d68290061de8e60376" Nov 25 22:15:30 crc kubenswrapper[4931]: E1125 22:15:30.117037 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-6bfcf864bd-cfmpv_openstack-operators(4e0590b1-c897-4382-bc93-4a1b0653a526)\"" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.121217 4931 generic.go:334] "Generic (PLEG): container finished" podID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" containerID="a1d3287876d55b59ee8bfd4cb6bc59ed106aeceefb23e987fd2878b261a986ac" exitCode=1 Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.121270 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" event={"ID":"60182b4c-52b4-47a9-8c5a-ab0274136ef1","Type":"ContainerDied","Data":"a1d3287876d55b59ee8bfd4cb6bc59ed106aeceefb23e987fd2878b261a986ac"} Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.121812 4931 scope.go:117] "RemoveContainer" containerID="a1d3287876d55b59ee8bfd4cb6bc59ed106aeceefb23e987fd2878b261a986ac" Nov 25 22:15:30 crc kubenswrapper[4931]: E1125 22:15:30.122197 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-5bfcf68b5b-79jjz_metallb-system(60182b4c-52b4-47a9-8c5a-ab0274136ef1)\"" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.135421 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.191651 4931 scope.go:117] "RemoveContainer" containerID="70c61a03853f7dba64993162bb7cd6128959f992ed3f36626f2bf1f5c54971e2" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.325679 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.381996 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.475269 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.475913 4931 scope.go:117] "RemoveContainer" containerID="f60a9d2362e3a3d4af899782755bda29a3f265b7103e5c9662a47be6301a5695" Nov 25 22:15:30 crc kubenswrapper[4931]: E1125 22:15:30.476137 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-6fb5bb54f-chfmg_openstack-operators(0700c2d1-0a31-4d46-9fe7-4bdb809d1cef)\"" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.487828 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.508042 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-jlksk" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.646159 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.851837 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pdm5c" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.863818 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.877945 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.891231 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.958047 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 22:15:30 crc kubenswrapper[4931]: I1125 22:15:30.980858 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.096892 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.124153 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-mw2wd" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.131789 4931 scope.go:117] "RemoveContainer" containerID="da4b6146230b9b514eef62c76d77f4e26423619e78cf37d68290061de8e60376" Nov 25 22:15:31 crc kubenswrapper[4931]: E1125 22:15:31.132291 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-6bfcf864bd-cfmpv_openstack-operators(4e0590b1-c897-4382-bc93-4a1b0653a526)\"" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.141869 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.165789 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.181889 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.222655 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.449457 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.463097 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.551152 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.680782 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.715767 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.775952 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.824797 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.855306 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.891344 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.918023 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.957193 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.960076 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.960752 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-6rxnn" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.961204 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.968455 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 22:15:31 crc kubenswrapper[4931]: I1125 22:15:31.971534 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.060119 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.094370 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-vn6l5" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.115225 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.115304 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.116240 4931 scope.go:117] "RemoveContainer" containerID="31ac46a90802ccc16582ea23119e002f6ae960d59b6b9deb90d377b0fafc85f1" Nov 25 22:15:32 crc kubenswrapper[4931]: E1125 22:15:32.116677 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-7b69997d9b-sz4qv_openstack-operators(952695c7-9eff-4e33-8777-6a5ea50a2b94)\"" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.150271 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.190510 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.214440 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.247334 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.254649 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.255784 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.294550 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.317858 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.432126 4931 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.471503 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.497700 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.498972 4931 scope.go:117] "RemoveContainer" containerID="28120008f68ab8447c744736d83346f31aae6356363b5efe208cea7a702c784d" Nov 25 22:15:32 crc kubenswrapper[4931]: E1125 22:15:32.499426 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-6545c6c9d6-7rkzq_openstack-operators(f4575978-a78e-42b2-a2dc-1052e5777b04)\"" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.539824 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.541529 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.585756 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.616725 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.625833 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 22:15:32 crc kubenswrapper[4931]: I1125 22:15:32.633225 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.026300 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-whzws" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.038577 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.089530 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.137396 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.137788 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.205098 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.219611 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.242184 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.287659 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.344177 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.393861 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.423649 4931 scope.go:117] "RemoveContainer" containerID="a4d5b21a4f10a37aa32a653d8f79b6eb0cabf19aaebf2e569325f4591a3cb8cd" Nov 25 22:15:33 crc kubenswrapper[4931]: E1125 22:15:33.424065 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-zsd98_openstack-operators(7b1bb711-ce82-4ea4-9667-ff4816b162f4)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" podUID="7b1bb711-ce82-4ea4-9667-ff4816b162f4" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.595827 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.598961 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.609386 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.647853 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.653928 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.692788 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.700262 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.716131 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.723021 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.765758 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.783681 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.891690 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.904478 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 22:15:33 crc kubenswrapper[4931]: I1125 22:15:33.930036 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.008457 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.097856 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.177769 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.203349 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.274355 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.369565 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.373379 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-2jtxv" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.376905 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.394978 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.483684 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.590183 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-fbp4q" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.593662 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 22:15:34 crc kubenswrapper[4931]: I1125 22:15:34.963076 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.036524 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.104023 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-9qv82" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.137789 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.173197 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.208985 4931 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.224929 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.227732 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.254187 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.254244 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.254916 4931 scope.go:117] "RemoveContainer" containerID="512092cb5bab7d49bb094278372ccd67706619dc39df920d362e11aec70170ba" Nov 25 22:15:35 crc kubenswrapper[4931]: E1125 22:15:35.255162 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-7cb66c858-rgrsz_openstack-operators(8f67c56d-e216-435c-8384-91d28a4d864e)\"" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" podUID="8f67c56d-e216-435c-8384-91d28a4d864e" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.257987 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.269659 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.283854 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.461168 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.531438 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.541960 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.573172 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.598478 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.690935 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.702452 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.714952 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.733684 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.737786 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.748036 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-lfhjt" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.861418 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.875148 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.892123 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-gvz2q" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.905015 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.954849 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 22:15:35 crc kubenswrapper[4931]: I1125 22:15:35.976120 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.010891 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.048600 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.065847 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.085481 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.258536 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.266833 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-pmmv2" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.269619 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.338143 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.345759 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.356797 4931 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.362562 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=40.362535854 podStartE2EDuration="40.362535854s" podCreationTimestamp="2025-11-25 22:14:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:15:15.86792468 +0000 UTC m=+1454.224436799" watchObservedRunningTime="2025-11-25 22:15:36.362535854 +0000 UTC m=+1474.719048003" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.362905 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.369509 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.369590 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.375994 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.378278 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.394075 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.394053342 podStartE2EDuration="21.394053342s" podCreationTimestamp="2025-11-25 22:15:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 22:15:36.386059807 +0000 UTC m=+1474.742571926" watchObservedRunningTime="2025-11-25 22:15:36.394053342 +0000 UTC m=+1474.750565471" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.489456 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.546699 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.547715 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.581075 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.603769 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.717788 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.737581 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.819551 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.831968 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.914080 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.915069 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.981902 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 22:15:36 crc kubenswrapper[4931]: I1125 22:15:36.991469 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.141984 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.154837 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.157604 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.164076 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.272643 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.288992 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.303980 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.304935 4931 scope.go:117] "RemoveContainer" containerID="7ba26f11563c6cc856c2d50ebc7125f560b7121a5d9f738fa872a9002aa33dda" Nov 25 22:15:37 crc kubenswrapper[4931]: E1125 22:15:37.305331 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-5b59697cf5-jthhc_openstack-operators(044c2639-0ceb-4be3-8326-4fdc552aefb6)\"" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.324668 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.389847 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.463802 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.612604 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.645265 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.645957 4931 scope.go:117] "RemoveContainer" containerID="a1d3287876d55b59ee8bfd4cb6bc59ed106aeceefb23e987fd2878b261a986ac" Nov 25 22:15:37 crc kubenswrapper[4931]: E1125 22:15:37.646203 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-5bfcf68b5b-79jjz_metallb-system(60182b4c-52b4-47a9-8c5a-ab0274136ef1)\"" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" podUID="60182b4c-52b4-47a9-8c5a-ab0274136ef1" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.665495 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.713788 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.720955 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.745519 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.826459 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.840087 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.853961 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.888264 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.914776 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.916967 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 22:15:37 crc kubenswrapper[4931]: I1125 22:15:37.967704 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.026780 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.039638 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.054494 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.062042 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.250446 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bpgdg" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.289499 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.292494 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.368101 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.404756 4931 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.404977 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f" gracePeriod=5 Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.409139 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.439838 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.499193 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.544825 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.586746 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.597310 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.608284 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.702657 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-5sjd4" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.723456 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.748507 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.801316 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.822233 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.844269 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.844337 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.844389 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.845349 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90db9e7682e94b8d54bbdd5d154989aeb94726c14a1c13fff2838dc83db574a2"} pod="openshift-machine-config-operator/machine-config-daemon-djjhg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.845422 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" containerID="cri-o://90db9e7682e94b8d54bbdd5d154989aeb94726c14a1c13fff2838dc83db574a2" gracePeriod=600 Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.887487 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd"] Nov 25 22:15:38 crc kubenswrapper[4931]: E1125 22:15:38.887851 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.887886 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 22:15:38 crc kubenswrapper[4931]: E1125 22:15:38.887904 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" containerName="installer" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.887912 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" containerName="installer" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.888076 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce336cf2-d251-4e2c-8879-2e7b48eaed18" containerName="installer" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.888109 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.888699 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.891744 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.892385 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.892801 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.910031 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd"] Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.915093 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.915508 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.933506 4931 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.963533 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.985130 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx2g2\" (UniqueName: \"kubernetes.io/projected/b427501e-c86c-4bcd-a698-7a56eed380ef-kube-api-access-cx2g2\") pod \"collect-profiles-29401815-dt8hd\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.985234 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b427501e-c86c-4bcd-a698-7a56eed380ef-config-volume\") pod \"collect-profiles-29401815-dt8hd\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:38 crc kubenswrapper[4931]: I1125 22:15:38.985253 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b427501e-c86c-4bcd-a698-7a56eed380ef-secret-volume\") pod \"collect-profiles-29401815-dt8hd\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.000089 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.030106 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.035244 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.086801 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b427501e-c86c-4bcd-a698-7a56eed380ef-config-volume\") pod \"collect-profiles-29401815-dt8hd\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.087176 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b427501e-c86c-4bcd-a698-7a56eed380ef-secret-volume\") pod \"collect-profiles-29401815-dt8hd\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.087403 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx2g2\" (UniqueName: \"kubernetes.io/projected/b427501e-c86c-4bcd-a698-7a56eed380ef-kube-api-access-cx2g2\") pod \"collect-profiles-29401815-dt8hd\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.088240 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b427501e-c86c-4bcd-a698-7a56eed380ef-config-volume\") pod \"collect-profiles-29401815-dt8hd\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.095923 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b427501e-c86c-4bcd-a698-7a56eed380ef-secret-volume\") pod \"collect-profiles-29401815-dt8hd\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.114570 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx2g2\" (UniqueName: \"kubernetes.io/projected/b427501e-c86c-4bcd-a698-7a56eed380ef-kube-api-access-cx2g2\") pod \"collect-profiles-29401815-dt8hd\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.184673 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.189904 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.204331 4931 generic.go:334] "Generic (PLEG): container finished" podID="e22448de-5122-43be-853a-081347c13cf2" containerID="90db9e7682e94b8d54bbdd5d154989aeb94726c14a1c13fff2838dc83db574a2" exitCode=0 Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.204411 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerDied","Data":"90db9e7682e94b8d54bbdd5d154989aeb94726c14a1c13fff2838dc83db574a2"} Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.204471 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerStarted","Data":"0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429"} Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.204490 4931 scope.go:117] "RemoveContainer" containerID="9ac1334287ef510421b58306fba1e2a55e64de3a45349db9fac90d16b42c433c" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.204973 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.205594 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.251707 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.262777 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.266453 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.341037 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.364017 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.414588 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.455161 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.561392 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.780289 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.835572 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.851186 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.861772 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.903218 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-b9jg9" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.944120 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.975996 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 22:15:39 crc kubenswrapper[4931]: I1125 22:15:39.998990 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.025521 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-gk9qz" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.114789 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.121248 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.124486 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.150432 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.233388 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.341260 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.438413 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 22:15:40 crc kubenswrapper[4931]: W1125 22:15:40.439326 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb427501e_c86c_4bcd_a698_7a56eed380ef.slice/crio-e966be981fa51229efe8b302cf32a4ecda240758196f98271764481ae3cc30da WatchSource:0}: Error finding container e966be981fa51229efe8b302cf32a4ecda240758196f98271764481ae3cc30da: Status 404 returned error can't find the container with id e966be981fa51229efe8b302cf32a4ecda240758196f98271764481ae3cc30da Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.446223 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.449709 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd"] Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.479241 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.479745 4931 scope.go:117] "RemoveContainer" containerID="f60a9d2362e3a3d4af899782755bda29a3f265b7103e5c9662a47be6301a5695" Nov 25 22:15:40 crc kubenswrapper[4931]: E1125 22:15:40.479979 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-6fb5bb54f-chfmg_openstack-operators(0700c2d1-0a31-4d46-9fe7-4bdb809d1cef)\"" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" podUID="0700c2d1-0a31-4d46-9fe7-4bdb809d1cef" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.484731 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.512152 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.586209 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-7wqhd" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.590802 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.734364 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.898610 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.953574 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-qsxk5" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.957118 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 22:15:40 crc kubenswrapper[4931]: I1125 22:15:40.976180 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.003001 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.033210 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.034787 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.101728 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.102672 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.166065 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-2j6qp" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.192283 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.193652 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.196828 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.240499 4931 generic.go:334] "Generic (PLEG): container finished" podID="b427501e-c86c-4bcd-a698-7a56eed380ef" containerID="a66dcb06bb80386a237a2398a05d5a627aaf809ea462d2f1de87254146c763ec" exitCode=0 Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.240544 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" event={"ID":"b427501e-c86c-4bcd-a698-7a56eed380ef","Type":"ContainerDied","Data":"a66dcb06bb80386a237a2398a05d5a627aaf809ea462d2f1de87254146c763ec"} Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.240570 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" event={"ID":"b427501e-c86c-4bcd-a698-7a56eed380ef","Type":"ContainerStarted","Data":"e966be981fa51229efe8b302cf32a4ecda240758196f98271764481ae3cc30da"} Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.243003 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.311186 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.475157 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.482159 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.502296 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.774292 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.828485 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.903074 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vd8m5" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.920499 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-zll5q" Nov 25 22:15:41 crc kubenswrapper[4931]: I1125 22:15:41.981581 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.006524 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.065954 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.161375 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.169461 4931 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.176440 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.194986 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.298251 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-wj9nf" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.341971 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.379093 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.409967 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.420030 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.465403 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.474163 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.496830 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.497773 4931 scope.go:117] "RemoveContainer" containerID="28120008f68ab8447c744736d83346f31aae6356363b5efe208cea7a702c784d" Nov 25 22:15:42 crc kubenswrapper[4931]: E1125 22:15:42.498169 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-6545c6c9d6-7rkzq_openstack-operators(f4575978-a78e-42b2-a2dc-1052e5777b04)\"" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" podUID="f4575978-a78e-42b2-a2dc-1052e5777b04" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.591144 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.646119 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b427501e-c86c-4bcd-a698-7a56eed380ef-config-volume\") pod \"b427501e-c86c-4bcd-a698-7a56eed380ef\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.646383 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b427501e-c86c-4bcd-a698-7a56eed380ef-secret-volume\") pod \"b427501e-c86c-4bcd-a698-7a56eed380ef\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.646551 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx2g2\" (UniqueName: \"kubernetes.io/projected/b427501e-c86c-4bcd-a698-7a56eed380ef-kube-api-access-cx2g2\") pod \"b427501e-c86c-4bcd-a698-7a56eed380ef\" (UID: \"b427501e-c86c-4bcd-a698-7a56eed380ef\") " Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.646816 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b427501e-c86c-4bcd-a698-7a56eed380ef-config-volume" (OuterVolumeSpecName: "config-volume") pod "b427501e-c86c-4bcd-a698-7a56eed380ef" (UID: "b427501e-c86c-4bcd-a698-7a56eed380ef"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.648024 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b427501e-c86c-4bcd-a698-7a56eed380ef-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.653450 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b427501e-c86c-4bcd-a698-7a56eed380ef-kube-api-access-cx2g2" (OuterVolumeSpecName: "kube-api-access-cx2g2") pod "b427501e-c86c-4bcd-a698-7a56eed380ef" (UID: "b427501e-c86c-4bcd-a698-7a56eed380ef"). InnerVolumeSpecName "kube-api-access-cx2g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.655771 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b427501e-c86c-4bcd-a698-7a56eed380ef-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b427501e-c86c-4bcd-a698-7a56eed380ef" (UID: "b427501e-c86c-4bcd-a698-7a56eed380ef"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.693503 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.749334 4931 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b427501e-c86c-4bcd-a698-7a56eed380ef-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.749387 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx2g2\" (UniqueName: \"kubernetes.io/projected/b427501e-c86c-4bcd-a698-7a56eed380ef-kube-api-access-cx2g2\") on node \"crc\" DevicePath \"\"" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.805967 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.929946 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.944734 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 22:15:42 crc kubenswrapper[4931]: I1125 22:15:42.997780 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.000362 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.037338 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.105667 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.113617 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.209299 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.264118 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" event={"ID":"b427501e-c86c-4bcd-a698-7a56eed380ef","Type":"ContainerDied","Data":"e966be981fa51229efe8b302cf32a4ecda240758196f98271764481ae3cc30da"} Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.264174 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e966be981fa51229efe8b302cf32a4ecda240758196f98271764481ae3cc30da" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.264255 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401815-dt8hd" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.398153 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.402278 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.594137 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.602371 4931 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.838852 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.891424 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-j5k42" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.920341 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 22:15:43 crc kubenswrapper[4931]: I1125 22:15:43.951095 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.038234 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.038335 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.071510 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.071569 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.071602 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.071608 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.071680 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.071740 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.071717 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.071730 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.071777 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.072135 4931 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.072155 4931 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.072167 4931 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.072179 4931 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.080602 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.174018 4931 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.178706 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.277834 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.277957 4931 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f" exitCode=137 Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.278020 4931 scope.go:117] "RemoveContainer" containerID="7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.278082 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.318003 4931 scope.go:117] "RemoveContainer" containerID="7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f" Nov 25 22:15:44 crc kubenswrapper[4931]: E1125 22:15:44.318750 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f\": container with ID starting with 7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f not found: ID does not exist" containerID="7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.318956 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f"} err="failed to get container status \"7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f\": rpc error: code = NotFound desc = could not find container \"7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f\": container with ID starting with 7ad73203198430867ae92d1148345a6f981045aba6b794f6ace63832e1a4449f not found: ID does not exist" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.334467 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.423292 4931 scope.go:117] "RemoveContainer" containerID="31ac46a90802ccc16582ea23119e002f6ae960d59b6b9deb90d377b0fafc85f1" Nov 25 22:15:44 crc kubenswrapper[4931]: E1125 22:15:44.423867 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-7b69997d9b-sz4qv_openstack-operators(952695c7-9eff-4e33-8777-6a5ea50a2b94)\"" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" podUID="952695c7-9eff-4e33-8777-6a5ea50a2b94" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.439635 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.440345 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.463153 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.463432 4931 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="32b67740-e63d-45ff-ab12-bf98a73dbb39" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.476111 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.476174 4931 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="32b67740-e63d-45ff-ab12-bf98a73dbb39" Nov 25 22:15:44 crc kubenswrapper[4931]: I1125 22:15:44.751384 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 22:15:45 crc kubenswrapper[4931]: I1125 22:15:45.290593 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 22:15:45 crc kubenswrapper[4931]: I1125 22:15:45.370236 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 22:15:45 crc kubenswrapper[4931]: I1125 22:15:45.423077 4931 scope.go:117] "RemoveContainer" containerID="a4d5b21a4f10a37aa32a653d8f79b6eb0cabf19aaebf2e569325f4591a3cb8cd" Nov 25 22:15:45 crc kubenswrapper[4931]: I1125 22:15:45.694787 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 22:15:46 crc kubenswrapper[4931]: I1125 22:15:46.298956 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-zsd98" event={"ID":"7b1bb711-ce82-4ea4-9667-ff4816b162f4","Type":"ContainerStarted","Data":"aefcee548191256b962c00b870ee4ff7029901e47d5b277078639e083eb0c07d"} Nov 25 22:15:46 crc kubenswrapper[4931]: I1125 22:15:46.421933 4931 scope.go:117] "RemoveContainer" containerID="da4b6146230b9b514eef62c76d77f4e26423619e78cf37d68290061de8e60376" Nov 25 22:15:46 crc kubenswrapper[4931]: E1125 22:15:46.422274 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-6bfcf864bd-cfmpv_openstack-operators(4e0590b1-c897-4382-bc93-4a1b0653a526)\"" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" Nov 25 22:15:46 crc kubenswrapper[4931]: I1125 22:15:46.422565 4931 scope.go:117] "RemoveContainer" containerID="512092cb5bab7d49bb094278372ccd67706619dc39df920d362e11aec70170ba" Nov 25 22:15:46 crc kubenswrapper[4931]: E1125 22:15:46.422953 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-7cb66c858-rgrsz_openstack-operators(8f67c56d-e216-435c-8384-91d28a4d864e)\"" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" podUID="8f67c56d-e216-435c-8384-91d28a4d864e" Nov 25 22:15:47 crc kubenswrapper[4931]: I1125 22:15:47.303603 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:15:47 crc kubenswrapper[4931]: I1125 22:15:47.306447 4931 scope.go:117] "RemoveContainer" containerID="7ba26f11563c6cc856c2d50ebc7125f560b7121a5d9f738fa872a9002aa33dda" Nov 25 22:15:47 crc kubenswrapper[4931]: E1125 22:15:47.307100 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-5b59697cf5-jthhc_openstack-operators(044c2639-0ceb-4be3-8326-4fdc552aefb6)\"" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" podUID="044c2639-0ceb-4be3-8326-4fdc552aefb6" Nov 25 22:15:48 crc kubenswrapper[4931]: I1125 22:15:48.728255 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:15:48 crc kubenswrapper[4931]: I1125 22:15:48.729237 4931 scope.go:117] "RemoveContainer" containerID="da4b6146230b9b514eef62c76d77f4e26423619e78cf37d68290061de8e60376" Nov 25 22:15:48 crc kubenswrapper[4931]: E1125 22:15:48.729534 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-6bfcf864bd-cfmpv_openstack-operators(4e0590b1-c897-4382-bc93-4a1b0653a526)\"" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" podUID="4e0590b1-c897-4382-bc93-4a1b0653a526" Nov 25 22:15:51 crc kubenswrapper[4931]: I1125 22:15:51.423000 4931 scope.go:117] "RemoveContainer" containerID="a1d3287876d55b59ee8bfd4cb6bc59ed106aeceefb23e987fd2878b261a986ac" Nov 25 22:15:52 crc kubenswrapper[4931]: I1125 22:15:52.367076 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" event={"ID":"60182b4c-52b4-47a9-8c5a-ab0274136ef1","Type":"ContainerStarted","Data":"5c910f4b6cf0d4dddd76d9b3b83fbfaf820d40e7995252ff064a1505eeff52b9"} Nov 25 22:15:52 crc kubenswrapper[4931]: I1125 22:15:52.367743 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:15:54 crc kubenswrapper[4931]: I1125 22:15:54.424107 4931 scope.go:117] "RemoveContainer" containerID="28120008f68ab8447c744736d83346f31aae6356363b5efe208cea7a702c784d" Nov 25 22:15:54 crc kubenswrapper[4931]: I1125 22:15:54.424978 4931 scope.go:117] "RemoveContainer" containerID="f60a9d2362e3a3d4af899782755bda29a3f265b7103e5c9662a47be6301a5695" Nov 25 22:15:55 crc kubenswrapper[4931]: I1125 22:15:55.401355 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" event={"ID":"f4575978-a78e-42b2-a2dc-1052e5777b04","Type":"ContainerStarted","Data":"0d8bcc3c2bc1810b0d28cb00c595f941623912c4aa36231c7384ea279dfdeff0"} Nov 25 22:15:55 crc kubenswrapper[4931]: I1125 22:15:55.402243 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:15:55 crc kubenswrapper[4931]: I1125 22:15:55.404213 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" event={"ID":"0700c2d1-0a31-4d46-9fe7-4bdb809d1cef","Type":"ContainerStarted","Data":"4177fcd495de1874a998e30278ad9b89a8f5693ec2b81d5c1c830ba4c8acbd8e"} Nov 25 22:15:55 crc kubenswrapper[4931]: I1125 22:15:55.404503 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:15:57 crc kubenswrapper[4931]: I1125 22:15:57.422358 4931 scope.go:117] "RemoveContainer" containerID="7ba26f11563c6cc856c2d50ebc7125f560b7121a5d9f738fa872a9002aa33dda" Nov 25 22:15:57 crc kubenswrapper[4931]: I1125 22:15:57.423180 4931 scope.go:117] "RemoveContainer" containerID="31ac46a90802ccc16582ea23119e002f6ae960d59b6b9deb90d377b0fafc85f1" Nov 25 22:15:58 crc kubenswrapper[4931]: I1125 22:15:58.449493 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" event={"ID":"044c2639-0ceb-4be3-8326-4fdc552aefb6","Type":"ContainerStarted","Data":"758b5819cb50253b425a97ac7db3aee0bfe75336494b19dd8264df8115ceb4bf"} Nov 25 22:15:58 crc kubenswrapper[4931]: I1125 22:15:58.449916 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:15:58 crc kubenswrapper[4931]: I1125 22:15:58.452396 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" event={"ID":"952695c7-9eff-4e33-8777-6a5ea50a2b94","Type":"ContainerStarted","Data":"7f493b13feb21cf5b7684e495c7fc65dec1f7a4be173c255129cf76d983929cf"} Nov 25 22:15:58 crc kubenswrapper[4931]: I1125 22:15:58.452585 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:15:59 crc kubenswrapper[4931]: I1125 22:15:59.423851 4931 scope.go:117] "RemoveContainer" containerID="512092cb5bab7d49bb094278372ccd67706619dc39df920d362e11aec70170ba" Nov 25 22:15:59 crc kubenswrapper[4931]: I1125 22:15:59.466337 4931 generic.go:334] "Generic (PLEG): container finished" podID="d0b65dd4-c1b3-4175-9951-3952c2fc4186" containerID="eb87536a2bcfa372ba61ea6efceb4d5debf9e912c0f80481fba7bc988fec041c" exitCode=0 Nov 25 22:15:59 crc kubenswrapper[4931]: I1125 22:15:59.466465 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" event={"ID":"d0b65dd4-c1b3-4175-9951-3952c2fc4186","Type":"ContainerDied","Data":"eb87536a2bcfa372ba61ea6efceb4d5debf9e912c0f80481fba7bc988fec041c"} Nov 25 22:15:59 crc kubenswrapper[4931]: I1125 22:15:59.467545 4931 scope.go:117] "RemoveContainer" containerID="eb87536a2bcfa372ba61ea6efceb4d5debf9e912c0f80481fba7bc988fec041c" Nov 25 22:16:00 crc kubenswrapper[4931]: I1125 22:16:00.482314 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-6fb5bb54f-chfmg" Nov 25 22:16:00 crc kubenswrapper[4931]: I1125 22:16:00.495016 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" event={"ID":"d0b65dd4-c1b3-4175-9951-3952c2fc4186","Type":"ContainerStarted","Data":"f24e244edccc3fa102d39746dcc7c53fc9bd8dd34d7b0dfa72908340221e09e3"} Nov 25 22:16:00 crc kubenswrapper[4931]: I1125 22:16:00.496340 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 22:16:00 crc kubenswrapper[4931]: I1125 22:16:00.499021 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" event={"ID":"8f67c56d-e216-435c-8384-91d28a4d864e","Type":"ContainerStarted","Data":"c67fe56ee2b3e106346883b6c986d49f1087a6233f8d91c1947774d03b4ec3f1"} Nov 25 22:16:00 crc kubenswrapper[4931]: I1125 22:16:00.499350 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:16:00 crc kubenswrapper[4931]: I1125 22:16:00.503925 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hffz8" Nov 25 22:16:02 crc kubenswrapper[4931]: I1125 22:16:02.120872 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7b69997d9b-sz4qv" Nov 25 22:16:02 crc kubenswrapper[4931]: I1125 22:16:02.439447 4931 scope.go:117] "RemoveContainer" containerID="da4b6146230b9b514eef62c76d77f4e26423619e78cf37d68290061de8e60376" Nov 25 22:16:02 crc kubenswrapper[4931]: I1125 22:16:02.501708 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6545c6c9d6-7rkzq" Nov 25 22:16:03 crc kubenswrapper[4931]: I1125 22:16:03.550063 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" event={"ID":"4e0590b1-c897-4382-bc93-4a1b0653a526","Type":"ContainerStarted","Data":"c6cd68dfc820262259ed29ec8d843143b787a450c2956b681aca3b0cba158929"} Nov 25 22:16:03 crc kubenswrapper[4931]: I1125 22:16:03.550762 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:16:03 crc kubenswrapper[4931]: I1125 22:16:03.779951 4931 scope.go:117] "RemoveContainer" containerID="3ca62f88382d274d24949e6f4a8a15e886ad177d4c690a3ac937f185773144e8" Nov 25 22:16:03 crc kubenswrapper[4931]: I1125 22:16:03.824187 4931 scope.go:117] "RemoveContainer" containerID="dc58e829c164e2f7e839ab52fafa5e3fca2e525d7a2172092ca55b8f7d30a019" Nov 25 22:16:03 crc kubenswrapper[4931]: I1125 22:16:03.868466 4931 scope.go:117] "RemoveContainer" containerID="956b3f180c7078f5ab4378ba41a8c5b80cb6bb0cd768ddbc26f371ae0260b20b" Nov 25 22:16:05 crc kubenswrapper[4931]: I1125 22:16:05.257670 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7cb66c858-rgrsz" Nov 25 22:16:07 crc kubenswrapper[4931]: I1125 22:16:07.312761 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5b59697cf5-jthhc" Nov 25 22:16:08 crc kubenswrapper[4931]: I1125 22:16:08.735771 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6bfcf864bd-cfmpv" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.512413 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nsr2j"] Nov 25 22:16:23 crc kubenswrapper[4931]: E1125 22:16:23.513490 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b427501e-c86c-4bcd-a698-7a56eed380ef" containerName="collect-profiles" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.513510 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b427501e-c86c-4bcd-a698-7a56eed380ef" containerName="collect-profiles" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.513766 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b427501e-c86c-4bcd-a698-7a56eed380ef" containerName="collect-profiles" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.515482 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.532726 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nsr2j"] Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.703441 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnbvp\" (UniqueName: \"kubernetes.io/projected/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-kube-api-access-lnbvp\") pod \"community-operators-nsr2j\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.703769 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-catalog-content\") pod \"community-operators-nsr2j\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.703909 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-utilities\") pod \"community-operators-nsr2j\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.805569 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnbvp\" (UniqueName: \"kubernetes.io/projected/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-kube-api-access-lnbvp\") pod \"community-operators-nsr2j\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.805636 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-catalog-content\") pod \"community-operators-nsr2j\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.805663 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-utilities\") pod \"community-operators-nsr2j\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.806308 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-catalog-content\") pod \"community-operators-nsr2j\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.806339 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-utilities\") pod \"community-operators-nsr2j\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.828184 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnbvp\" (UniqueName: \"kubernetes.io/projected/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-kube-api-access-lnbvp\") pod \"community-operators-nsr2j\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:23 crc kubenswrapper[4931]: I1125 22:16:23.835930 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:24 crc kubenswrapper[4931]: I1125 22:16:24.333169 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nsr2j"] Nov 25 22:16:24 crc kubenswrapper[4931]: W1125 22:16:24.335695 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4ab7eb9_dc46_4513_8dec_f16ff003c5d8.slice/crio-6db512dddfb0943377fabe06fbb8ba35a0a33116153d68eb1cbdfb81855aef7f WatchSource:0}: Error finding container 6db512dddfb0943377fabe06fbb8ba35a0a33116153d68eb1cbdfb81855aef7f: Status 404 returned error can't find the container with id 6db512dddfb0943377fabe06fbb8ba35a0a33116153d68eb1cbdfb81855aef7f Nov 25 22:16:24 crc kubenswrapper[4931]: I1125 22:16:24.721466 4931 generic.go:334] "Generic (PLEG): container finished" podID="a4ab7eb9-dc46-4513-8dec-f16ff003c5d8" containerID="ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd" exitCode=0 Nov 25 22:16:24 crc kubenswrapper[4931]: I1125 22:16:24.721552 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsr2j" event={"ID":"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8","Type":"ContainerDied","Data":"ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd"} Nov 25 22:16:24 crc kubenswrapper[4931]: I1125 22:16:24.721727 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsr2j" event={"ID":"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8","Type":"ContainerStarted","Data":"6db512dddfb0943377fabe06fbb8ba35a0a33116153d68eb1cbdfb81855aef7f"} Nov 25 22:16:24 crc kubenswrapper[4931]: I1125 22:16:24.723978 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 22:16:25 crc kubenswrapper[4931]: I1125 22:16:25.734312 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsr2j" event={"ID":"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8","Type":"ContainerStarted","Data":"36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434"} Nov 25 22:16:26 crc kubenswrapper[4931]: I1125 22:16:26.746296 4931 generic.go:334] "Generic (PLEG): container finished" podID="a4ab7eb9-dc46-4513-8dec-f16ff003c5d8" containerID="36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434" exitCode=0 Nov 25 22:16:26 crc kubenswrapper[4931]: I1125 22:16:26.746384 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsr2j" event={"ID":"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8","Type":"ContainerDied","Data":"36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434"} Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.311948 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bfggt"] Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.313965 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.330387 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfggt"] Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.362950 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lsqs\" (UniqueName: \"kubernetes.io/projected/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-kube-api-access-2lsqs\") pod \"redhat-marketplace-bfggt\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.369204 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-utilities\") pod \"redhat-marketplace-bfggt\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.369633 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-catalog-content\") pod \"redhat-marketplace-bfggt\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.470981 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lsqs\" (UniqueName: \"kubernetes.io/projected/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-kube-api-access-2lsqs\") pod \"redhat-marketplace-bfggt\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.471512 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-utilities\") pod \"redhat-marketplace-bfggt\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.471660 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-catalog-content\") pod \"redhat-marketplace-bfggt\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.472484 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-utilities\") pod \"redhat-marketplace-bfggt\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.472788 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-catalog-content\") pod \"redhat-marketplace-bfggt\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.500094 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lsqs\" (UniqueName: \"kubernetes.io/projected/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-kube-api-access-2lsqs\") pod \"redhat-marketplace-bfggt\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.646681 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5bfcf68b5b-79jjz" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.648369 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.763439 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsr2j" event={"ID":"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8","Type":"ContainerStarted","Data":"80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9"} Nov 25 22:16:27 crc kubenswrapper[4931]: I1125 22:16:27.790228 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nsr2j" podStartSLOduration=2.223245449 podStartE2EDuration="4.790212524s" podCreationTimestamp="2025-11-25 22:16:23 +0000 UTC" firstStartedPulling="2025-11-25 22:16:24.723676582 +0000 UTC m=+1523.080188711" lastFinishedPulling="2025-11-25 22:16:27.290643657 +0000 UTC m=+1525.647155786" observedRunningTime="2025-11-25 22:16:27.788104947 +0000 UTC m=+1526.144617066" watchObservedRunningTime="2025-11-25 22:16:27.790212524 +0000 UTC m=+1526.146724643" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.149205 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfggt"] Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.309783 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kjrgw"] Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.311382 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.327811 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kjrgw"] Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.383231 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-utilities\") pod \"redhat-operators-kjrgw\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.383265 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq4m8\" (UniqueName: \"kubernetes.io/projected/a441c870-fc75-42d3-9c5a-2e432913db7f-kube-api-access-rq4m8\") pod \"redhat-operators-kjrgw\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.383333 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-catalog-content\") pod \"redhat-operators-kjrgw\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.484600 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-catalog-content\") pod \"redhat-operators-kjrgw\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.484688 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-utilities\") pod \"redhat-operators-kjrgw\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.484708 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq4m8\" (UniqueName: \"kubernetes.io/projected/a441c870-fc75-42d3-9c5a-2e432913db7f-kube-api-access-rq4m8\") pod \"redhat-operators-kjrgw\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.485176 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-utilities\") pod \"redhat-operators-kjrgw\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.485511 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-catalog-content\") pod \"redhat-operators-kjrgw\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.501765 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq4m8\" (UniqueName: \"kubernetes.io/projected/a441c870-fc75-42d3-9c5a-2e432913db7f-kube-api-access-rq4m8\") pod \"redhat-operators-kjrgw\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.626109 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.813467 4931 generic.go:334] "Generic (PLEG): container finished" podID="3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862" containerID="5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98" exitCode=0 Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.814696 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfggt" event={"ID":"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862","Type":"ContainerDied","Data":"5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98"} Nov 25 22:16:28 crc kubenswrapper[4931]: I1125 22:16:28.814721 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfggt" event={"ID":"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862","Type":"ContainerStarted","Data":"d52ca52c2a7db501a8d9894b210f39962ece48ce65c98b500cd27e46325e513a"} Nov 25 22:16:29 crc kubenswrapper[4931]: I1125 22:16:29.141832 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kjrgw"] Nov 25 22:16:29 crc kubenswrapper[4931]: W1125 22:16:29.150063 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda441c870_fc75_42d3_9c5a_2e432913db7f.slice/crio-f95ed2c006535122e18e01c1e4989525c10d62819454a09e206d7f791b0d732b WatchSource:0}: Error finding container f95ed2c006535122e18e01c1e4989525c10d62819454a09e206d7f791b0d732b: Status 404 returned error can't find the container with id f95ed2c006535122e18e01c1e4989525c10d62819454a09e206d7f791b0d732b Nov 25 22:16:29 crc kubenswrapper[4931]: I1125 22:16:29.822094 4931 generic.go:334] "Generic (PLEG): container finished" podID="3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862" containerID="f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7" exitCode=0 Nov 25 22:16:29 crc kubenswrapper[4931]: I1125 22:16:29.822165 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfggt" event={"ID":"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862","Type":"ContainerDied","Data":"f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7"} Nov 25 22:16:29 crc kubenswrapper[4931]: I1125 22:16:29.823966 4931 generic.go:334] "Generic (PLEG): container finished" podID="a441c870-fc75-42d3-9c5a-2e432913db7f" containerID="70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e" exitCode=0 Nov 25 22:16:29 crc kubenswrapper[4931]: I1125 22:16:29.823998 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kjrgw" event={"ID":"a441c870-fc75-42d3-9c5a-2e432913db7f","Type":"ContainerDied","Data":"70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e"} Nov 25 22:16:29 crc kubenswrapper[4931]: I1125 22:16:29.824018 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kjrgw" event={"ID":"a441c870-fc75-42d3-9c5a-2e432913db7f","Type":"ContainerStarted","Data":"f95ed2c006535122e18e01c1e4989525c10d62819454a09e206d7f791b0d732b"} Nov 25 22:16:30 crc kubenswrapper[4931]: I1125 22:16:30.834048 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfggt" event={"ID":"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862","Type":"ContainerStarted","Data":"aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9"} Nov 25 22:16:31 crc kubenswrapper[4931]: I1125 22:16:31.861695 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kjrgw" event={"ID":"a441c870-fc75-42d3-9c5a-2e432913db7f","Type":"ContainerStarted","Data":"951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b"} Nov 25 22:16:31 crc kubenswrapper[4931]: I1125 22:16:31.890983 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bfggt" podStartSLOduration=3.497546127 podStartE2EDuration="4.890965744s" podCreationTimestamp="2025-11-25 22:16:27 +0000 UTC" firstStartedPulling="2025-11-25 22:16:28.817335521 +0000 UTC m=+1527.173847640" lastFinishedPulling="2025-11-25 22:16:30.210755138 +0000 UTC m=+1528.567267257" observedRunningTime="2025-11-25 22:16:30.857242869 +0000 UTC m=+1529.213754978" watchObservedRunningTime="2025-11-25 22:16:31.890965744 +0000 UTC m=+1530.247477863" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.680202 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ksnv8/must-gather-hbhpw"] Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.682145 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.683855 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ksnv8"/"kube-root-ca.crt" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.684588 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ksnv8"/"openshift-service-ca.crt" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.685332 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ksnv8"/"default-dockercfg-vltdh" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.690661 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ksnv8/must-gather-hbhpw"] Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.782267 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xwx2\" (UniqueName: \"kubernetes.io/projected/aea87457-cffd-4a2d-a368-4c5efb5941ac-kube-api-access-9xwx2\") pod \"must-gather-hbhpw\" (UID: \"aea87457-cffd-4a2d-a368-4c5efb5941ac\") " pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.782546 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aea87457-cffd-4a2d-a368-4c5efb5941ac-must-gather-output\") pod \"must-gather-hbhpw\" (UID: \"aea87457-cffd-4a2d-a368-4c5efb5941ac\") " pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.836738 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.836923 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.879340 4931 generic.go:334] "Generic (PLEG): container finished" podID="a441c870-fc75-42d3-9c5a-2e432913db7f" containerID="951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b" exitCode=0 Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.879591 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kjrgw" event={"ID":"a441c870-fc75-42d3-9c5a-2e432913db7f","Type":"ContainerDied","Data":"951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b"} Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.884826 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xwx2\" (UniqueName: \"kubernetes.io/projected/aea87457-cffd-4a2d-a368-4c5efb5941ac-kube-api-access-9xwx2\") pod \"must-gather-hbhpw\" (UID: \"aea87457-cffd-4a2d-a368-4c5efb5941ac\") " pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.884959 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aea87457-cffd-4a2d-a368-4c5efb5941ac-must-gather-output\") pod \"must-gather-hbhpw\" (UID: \"aea87457-cffd-4a2d-a368-4c5efb5941ac\") " pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.885488 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aea87457-cffd-4a2d-a368-4c5efb5941ac-must-gather-output\") pod \"must-gather-hbhpw\" (UID: \"aea87457-cffd-4a2d-a368-4c5efb5941ac\") " pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.889310 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:33 crc kubenswrapper[4931]: I1125 22:16:33.919473 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xwx2\" (UniqueName: \"kubernetes.io/projected/aea87457-cffd-4a2d-a368-4c5efb5941ac-kube-api-access-9xwx2\") pod \"must-gather-hbhpw\" (UID: \"aea87457-cffd-4a2d-a368-4c5efb5941ac\") " pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:16:34 crc kubenswrapper[4931]: I1125 22:16:34.077249 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:16:34 crc kubenswrapper[4931]: I1125 22:16:34.584367 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ksnv8/must-gather-hbhpw"] Nov 25 22:16:34 crc kubenswrapper[4931]: W1125 22:16:34.594181 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaea87457_cffd_4a2d_a368_4c5efb5941ac.slice/crio-4519e09b31a0ef5f625d52a2b2ea51890b48de7de176de750bf4242f7996d2af WatchSource:0}: Error finding container 4519e09b31a0ef5f625d52a2b2ea51890b48de7de176de750bf4242f7996d2af: Status 404 returned error can't find the container with id 4519e09b31a0ef5f625d52a2b2ea51890b48de7de176de750bf4242f7996d2af Nov 25 22:16:34 crc kubenswrapper[4931]: I1125 22:16:34.888195 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" event={"ID":"aea87457-cffd-4a2d-a368-4c5efb5941ac","Type":"ContainerStarted","Data":"4519e09b31a0ef5f625d52a2b2ea51890b48de7de176de750bf4242f7996d2af"} Nov 25 22:16:34 crc kubenswrapper[4931]: I1125 22:16:34.951837 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:35 crc kubenswrapper[4931]: I1125 22:16:35.898897 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kjrgw" event={"ID":"a441c870-fc75-42d3-9c5a-2e432913db7f","Type":"ContainerStarted","Data":"e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10"} Nov 25 22:16:37 crc kubenswrapper[4931]: I1125 22:16:37.648553 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:37 crc kubenswrapper[4931]: I1125 22:16:37.648929 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:37 crc kubenswrapper[4931]: I1125 22:16:37.697856 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:37 crc kubenswrapper[4931]: I1125 22:16:37.717507 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kjrgw" podStartSLOduration=4.52201795 podStartE2EDuration="9.717483617s" podCreationTimestamp="2025-11-25 22:16:28 +0000 UTC" firstStartedPulling="2025-11-25 22:16:29.825323703 +0000 UTC m=+1528.181835822" lastFinishedPulling="2025-11-25 22:16:35.02078937 +0000 UTC m=+1533.377301489" observedRunningTime="2025-11-25 22:16:35.918239576 +0000 UTC m=+1534.274751695" watchObservedRunningTime="2025-11-25 22:16:37.717483617 +0000 UTC m=+1536.073995736" Nov 25 22:16:37 crc kubenswrapper[4931]: I1125 22:16:37.909941 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k2bjv"] Nov 25 22:16:37 crc kubenswrapper[4931]: I1125 22:16:37.912679 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:37 crc kubenswrapper[4931]: I1125 22:16:37.925698 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k2bjv"] Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.031650 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.056716 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-catalog-content\") pod \"certified-operators-k2bjv\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.056915 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9slw\" (UniqueName: \"kubernetes.io/projected/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-kube-api-access-v9slw\") pod \"certified-operators-k2bjv\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.057012 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-utilities\") pod \"certified-operators-k2bjv\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.164653 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-catalog-content\") pod \"certified-operators-k2bjv\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.165040 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9slw\" (UniqueName: \"kubernetes.io/projected/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-kube-api-access-v9slw\") pod \"certified-operators-k2bjv\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.165117 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-utilities\") pod \"certified-operators-k2bjv\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.165431 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-catalog-content\") pod \"certified-operators-k2bjv\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.165819 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-utilities\") pod \"certified-operators-k2bjv\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.189539 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9slw\" (UniqueName: \"kubernetes.io/projected/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-kube-api-access-v9slw\") pod \"certified-operators-k2bjv\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.278016 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.627474 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:38 crc kubenswrapper[4931]: I1125 22:16:38.627819 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:39 crc kubenswrapper[4931]: I1125 22:16:39.673651 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kjrgw" podUID="a441c870-fc75-42d3-9c5a-2e432913db7f" containerName="registry-server" probeResult="failure" output=< Nov 25 22:16:39 crc kubenswrapper[4931]: timeout: failed to connect service ":50051" within 1s Nov 25 22:16:39 crc kubenswrapper[4931]: > Nov 25 22:16:39 crc kubenswrapper[4931]: I1125 22:16:39.936418 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" event={"ID":"aea87457-cffd-4a2d-a368-4c5efb5941ac","Type":"ContainerStarted","Data":"87a8dc8ef076c36dbd3dd63f8f5d2fbc9255f32469bbb50df6a85c78ae825d70"} Nov 25 22:16:40 crc kubenswrapper[4931]: I1125 22:16:40.070254 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k2bjv"] Nov 25 22:16:40 crc kubenswrapper[4931]: I1125 22:16:40.949845 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" event={"ID":"aea87457-cffd-4a2d-a368-4c5efb5941ac","Type":"ContainerStarted","Data":"8ff403f34d8bc6d8bca03a3041c257199bf830ac1d3815cc785d53c0d8d92fdd"} Nov 25 22:16:40 crc kubenswrapper[4931]: I1125 22:16:40.951992 4931 generic.go:334] "Generic (PLEG): container finished" podID="5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e" containerID="5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0" exitCode=0 Nov 25 22:16:40 crc kubenswrapper[4931]: I1125 22:16:40.952092 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k2bjv" event={"ID":"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e","Type":"ContainerDied","Data":"5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0"} Nov 25 22:16:40 crc kubenswrapper[4931]: I1125 22:16:40.952157 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k2bjv" event={"ID":"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e","Type":"ContainerStarted","Data":"bb384fe53a307c11997a02e8c2119a99b276c40fd36d9db8055f8c14e67504db"} Nov 25 22:16:40 crc kubenswrapper[4931]: I1125 22:16:40.964635 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" podStartSLOduration=2.918575166 podStartE2EDuration="7.964619331s" podCreationTimestamp="2025-11-25 22:16:33 +0000 UTC" firstStartedPulling="2025-11-25 22:16:34.595622076 +0000 UTC m=+1532.952134205" lastFinishedPulling="2025-11-25 22:16:39.641666241 +0000 UTC m=+1537.998178370" observedRunningTime="2025-11-25 22:16:40.962474812 +0000 UTC m=+1539.318986941" watchObservedRunningTime="2025-11-25 22:16:40.964619331 +0000 UTC m=+1539.321131450" Nov 25 22:16:41 crc kubenswrapper[4931]: I1125 22:16:41.962642 4931 generic.go:334] "Generic (PLEG): container finished" podID="5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e" containerID="d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007" exitCode=0 Nov 25 22:16:41 crc kubenswrapper[4931]: I1125 22:16:41.963431 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k2bjv" event={"ID":"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e","Type":"ContainerDied","Data":"d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007"} Nov 25 22:16:43 crc kubenswrapper[4931]: I1125 22:16:43.979663 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k2bjv" event={"ID":"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e","Type":"ContainerStarted","Data":"b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819"} Nov 25 22:16:44 crc kubenswrapper[4931]: I1125 22:16:44.007100 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k2bjv" podStartSLOduration=4.416372141 podStartE2EDuration="7.007077875s" podCreationTimestamp="2025-11-25 22:16:37 +0000 UTC" firstStartedPulling="2025-11-25 22:16:40.954062066 +0000 UTC m=+1539.310574195" lastFinishedPulling="2025-11-25 22:16:43.54476781 +0000 UTC m=+1541.901279929" observedRunningTime="2025-11-25 22:16:44.003289243 +0000 UTC m=+1542.359801362" watchObservedRunningTime="2025-11-25 22:16:44.007077875 +0000 UTC m=+1542.363590034" Nov 25 22:16:48 crc kubenswrapper[4931]: I1125 22:16:48.279104 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:48 crc kubenswrapper[4931]: I1125 22:16:48.279901 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:48 crc kubenswrapper[4931]: I1125 22:16:48.336289 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:48 crc kubenswrapper[4931]: I1125 22:16:48.672462 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:48 crc kubenswrapper[4931]: I1125 22:16:48.727802 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:16:49 crc kubenswrapper[4931]: I1125 22:16:49.073285 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.301187 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k2bjv"] Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.303532 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k2bjv" podUID="5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e" containerName="registry-server" containerID="cri-o://b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819" gracePeriod=2 Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.497136 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nsr2j"] Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.497424 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nsr2j" podUID="a4ab7eb9-dc46-4513-8dec-f16ff003c5d8" containerName="registry-server" containerID="cri-o://80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9" gracePeriod=2 Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.752057 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.874030 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-catalog-content\") pod \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.874469 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-utilities\") pod \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.874516 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9slw\" (UniqueName: \"kubernetes.io/projected/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-kube-api-access-v9slw\") pod \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\" (UID: \"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e\") " Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.875350 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-utilities" (OuterVolumeSpecName: "utilities") pod "5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e" (UID: "5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.881245 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-kube-api-access-v9slw" (OuterVolumeSpecName: "kube-api-access-v9slw") pod "5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e" (UID: "5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e"). InnerVolumeSpecName "kube-api-access-v9slw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.921974 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e" (UID: "5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.930989 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.976150 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.976189 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:16:55 crc kubenswrapper[4931]: I1125 22:16:55.976203 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9slw\" (UniqueName: \"kubernetes.io/projected/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e-kube-api-access-v9slw\") on node \"crc\" DevicePath \"\"" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.076637 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnbvp\" (UniqueName: \"kubernetes.io/projected/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-kube-api-access-lnbvp\") pod \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.076964 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-utilities\") pod \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.077008 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-catalog-content\") pod \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\" (UID: \"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8\") " Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.078262 4931 generic.go:334] "Generic (PLEG): container finished" podID="a4ab7eb9-dc46-4513-8dec-f16ff003c5d8" containerID="80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9" exitCode=0 Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.078310 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsr2j" event={"ID":"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8","Type":"ContainerDied","Data":"80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9"} Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.078335 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsr2j" event={"ID":"a4ab7eb9-dc46-4513-8dec-f16ff003c5d8","Type":"ContainerDied","Data":"6db512dddfb0943377fabe06fbb8ba35a0a33116153d68eb1cbdfb81855aef7f"} Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.078352 4931 scope.go:117] "RemoveContainer" containerID="80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.078450 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsr2j" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.078496 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-utilities" (OuterVolumeSpecName: "utilities") pod "a4ab7eb9-dc46-4513-8dec-f16ff003c5d8" (UID: "a4ab7eb9-dc46-4513-8dec-f16ff003c5d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.086051 4931 generic.go:334] "Generic (PLEG): container finished" podID="5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e" containerID="b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819" exitCode=0 Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.086095 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k2bjv" event={"ID":"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e","Type":"ContainerDied","Data":"b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819"} Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.086125 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k2bjv" event={"ID":"5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e","Type":"ContainerDied","Data":"bb384fe53a307c11997a02e8c2119a99b276c40fd36d9db8055f8c14e67504db"} Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.086081 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-kube-api-access-lnbvp" (OuterVolumeSpecName: "kube-api-access-lnbvp") pod "a4ab7eb9-dc46-4513-8dec-f16ff003c5d8" (UID: "a4ab7eb9-dc46-4513-8dec-f16ff003c5d8"). InnerVolumeSpecName "kube-api-access-lnbvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.086253 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k2bjv" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.100319 4931 scope.go:117] "RemoveContainer" containerID="36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.123827 4931 scope.go:117] "RemoveContainer" containerID="ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.126310 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k2bjv"] Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.132367 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4ab7eb9-dc46-4513-8dec-f16ff003c5d8" (UID: "a4ab7eb9-dc46-4513-8dec-f16ff003c5d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.136154 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k2bjv"] Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.142125 4931 scope.go:117] "RemoveContainer" containerID="80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9" Nov 25 22:16:56 crc kubenswrapper[4931]: E1125 22:16:56.142612 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9\": container with ID starting with 80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9 not found: ID does not exist" containerID="80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.142649 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9"} err="failed to get container status \"80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9\": rpc error: code = NotFound desc = could not find container \"80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9\": container with ID starting with 80230544498e3a1f3b8e2dacee1c187aab61e418664af00393482ed8b17cedb9 not found: ID does not exist" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.142693 4931 scope.go:117] "RemoveContainer" containerID="36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434" Nov 25 22:16:56 crc kubenswrapper[4931]: E1125 22:16:56.143092 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434\": container with ID starting with 36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434 not found: ID does not exist" containerID="36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.143117 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434"} err="failed to get container status \"36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434\": rpc error: code = NotFound desc = could not find container \"36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434\": container with ID starting with 36d9d1d61a8502152c44b52e2fa7b6b4f9339ae9dfef59d16dc3840e64349434 not found: ID does not exist" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.143134 4931 scope.go:117] "RemoveContainer" containerID="ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd" Nov 25 22:16:56 crc kubenswrapper[4931]: E1125 22:16:56.143392 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd\": container with ID starting with ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd not found: ID does not exist" containerID="ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.143414 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd"} err="failed to get container status \"ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd\": rpc error: code = NotFound desc = could not find container \"ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd\": container with ID starting with ecf7dfa181d92302d33f70c64079fefb0cf09583d7f528ad8b2a849fe95c07dd not found: ID does not exist" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.143429 4931 scope.go:117] "RemoveContainer" containerID="b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.167276 4931 scope.go:117] "RemoveContainer" containerID="d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.179509 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.179537 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.179546 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnbvp\" (UniqueName: \"kubernetes.io/projected/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8-kube-api-access-lnbvp\") on node \"crc\" DevicePath \"\"" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.194968 4931 scope.go:117] "RemoveContainer" containerID="5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.214752 4931 scope.go:117] "RemoveContainer" containerID="b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819" Nov 25 22:16:56 crc kubenswrapper[4931]: E1125 22:16:56.215202 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819\": container with ID starting with b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819 not found: ID does not exist" containerID="b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.215231 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819"} err="failed to get container status \"b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819\": rpc error: code = NotFound desc = could not find container \"b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819\": container with ID starting with b57e419524933b90a96f8a4b4103121d5934d2c00ac5a36e5d4b390e3d661819 not found: ID does not exist" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.215252 4931 scope.go:117] "RemoveContainer" containerID="d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007" Nov 25 22:16:56 crc kubenswrapper[4931]: E1125 22:16:56.215573 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007\": container with ID starting with d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007 not found: ID does not exist" containerID="d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.215596 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007"} err="failed to get container status \"d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007\": rpc error: code = NotFound desc = could not find container \"d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007\": container with ID starting with d2f39642ed225295affb5e8ddbd364e6cd55f0de2848f008ebd5da839648f007 not found: ID does not exist" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.215612 4931 scope.go:117] "RemoveContainer" containerID="5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0" Nov 25 22:16:56 crc kubenswrapper[4931]: E1125 22:16:56.215832 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0\": container with ID starting with 5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0 not found: ID does not exist" containerID="5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.215852 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0"} err="failed to get container status \"5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0\": rpc error: code = NotFound desc = could not find container \"5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0\": container with ID starting with 5d98271279785c657fdbfe390482dbb1fed70e91c6df2803bcceda4c008a97c0 not found: ID does not exist" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.420123 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nsr2j"] Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.432321 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e" path="/var/lib/kubelet/pods/5ec80460-7073-4ab4-bdb7-5ee37dc9ba5e/volumes" Nov 25 22:16:56 crc kubenswrapper[4931]: I1125 22:16:56.433392 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nsr2j"] Nov 25 22:16:58 crc kubenswrapper[4931]: I1125 22:16:58.432065 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4ab7eb9-dc46-4513-8dec-f16ff003c5d8" path="/var/lib/kubelet/pods/a4ab7eb9-dc46-4513-8dec-f16ff003c5d8/volumes" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.297504 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfggt"] Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.298192 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bfggt" podUID="3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862" containerName="registry-server" containerID="cri-o://aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9" gracePeriod=2 Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.492909 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kjrgw"] Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.493447 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kjrgw" podUID="a441c870-fc75-42d3-9c5a-2e432913db7f" containerName="registry-server" containerID="cri-o://e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10" gracePeriod=2 Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.741411 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.853160 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-utilities\") pod \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.853600 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-catalog-content\") pod \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.853707 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lsqs\" (UniqueName: \"kubernetes.io/projected/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-kube-api-access-2lsqs\") pod \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\" (UID: \"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862\") " Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.853888 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-utilities" (OuterVolumeSpecName: "utilities") pod "3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862" (UID: "3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.854241 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.864021 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-kube-api-access-2lsqs" (OuterVolumeSpecName: "kube-api-access-2lsqs") pod "3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862" (UID: "3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862"). InnerVolumeSpecName "kube-api-access-2lsqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.868401 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862" (UID: "3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.884614 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.954730 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq4m8\" (UniqueName: \"kubernetes.io/projected/a441c870-fc75-42d3-9c5a-2e432913db7f-kube-api-access-rq4m8\") pod \"a441c870-fc75-42d3-9c5a-2e432913db7f\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.954863 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-utilities\") pod \"a441c870-fc75-42d3-9c5a-2e432913db7f\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.954972 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-catalog-content\") pod \"a441c870-fc75-42d3-9c5a-2e432913db7f\" (UID: \"a441c870-fc75-42d3-9c5a-2e432913db7f\") " Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.955261 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lsqs\" (UniqueName: \"kubernetes.io/projected/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-kube-api-access-2lsqs\") on node \"crc\" DevicePath \"\"" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.955273 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.956984 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-utilities" (OuterVolumeSpecName: "utilities") pod "a441c870-fc75-42d3-9c5a-2e432913db7f" (UID: "a441c870-fc75-42d3-9c5a-2e432913db7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:17:00 crc kubenswrapper[4931]: I1125 22:17:00.958245 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a441c870-fc75-42d3-9c5a-2e432913db7f-kube-api-access-rq4m8" (OuterVolumeSpecName: "kube-api-access-rq4m8") pod "a441c870-fc75-42d3-9c5a-2e432913db7f" (UID: "a441c870-fc75-42d3-9c5a-2e432913db7f"). InnerVolumeSpecName "kube-api-access-rq4m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.026316 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a441c870-fc75-42d3-9c5a-2e432913db7f" (UID: "a441c870-fc75-42d3-9c5a-2e432913db7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.056912 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.056942 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq4m8\" (UniqueName: \"kubernetes.io/projected/a441c870-fc75-42d3-9c5a-2e432913db7f-kube-api-access-rq4m8\") on node \"crc\" DevicePath \"\"" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.056952 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a441c870-fc75-42d3-9c5a-2e432913db7f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.133703 4931 generic.go:334] "Generic (PLEG): container finished" podID="a441c870-fc75-42d3-9c5a-2e432913db7f" containerID="e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10" exitCode=0 Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.133771 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kjrgw" event={"ID":"a441c870-fc75-42d3-9c5a-2e432913db7f","Type":"ContainerDied","Data":"e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10"} Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.133783 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kjrgw" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.133813 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kjrgw" event={"ID":"a441c870-fc75-42d3-9c5a-2e432913db7f","Type":"ContainerDied","Data":"f95ed2c006535122e18e01c1e4989525c10d62819454a09e206d7f791b0d732b"} Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.133835 4931 scope.go:117] "RemoveContainer" containerID="e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.137126 4931 generic.go:334] "Generic (PLEG): container finished" podID="3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862" containerID="aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9" exitCode=0 Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.137154 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfggt" event={"ID":"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862","Type":"ContainerDied","Data":"aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9"} Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.137194 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfggt" event={"ID":"3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862","Type":"ContainerDied","Data":"d52ca52c2a7db501a8d9894b210f39962ece48ce65c98b500cd27e46325e513a"} Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.137271 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfggt" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.155000 4931 scope.go:117] "RemoveContainer" containerID="951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.173585 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfggt"] Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.188524 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfggt"] Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.191133 4931 scope.go:117] "RemoveContainer" containerID="70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.199933 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kjrgw"] Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.206951 4931 scope.go:117] "RemoveContainer" containerID="e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10" Nov 25 22:17:01 crc kubenswrapper[4931]: E1125 22:17:01.207331 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10\": container with ID starting with e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10 not found: ID does not exist" containerID="e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.207372 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10"} err="failed to get container status \"e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10\": rpc error: code = NotFound desc = could not find container \"e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10\": container with ID starting with e738675e3c58258f9845f045bca1b6360102a83499264a5ec19d76c0af490e10 not found: ID does not exist" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.207394 4931 scope.go:117] "RemoveContainer" containerID="951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b" Nov 25 22:17:01 crc kubenswrapper[4931]: E1125 22:17:01.207721 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b\": container with ID starting with 951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b not found: ID does not exist" containerID="951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.207744 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b"} err="failed to get container status \"951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b\": rpc error: code = NotFound desc = could not find container \"951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b\": container with ID starting with 951dfdf85604aed1b5b7854d6070425cdf7031715c22a67595dc47de9c11c09b not found: ID does not exist" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.207757 4931 scope.go:117] "RemoveContainer" containerID="70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e" Nov 25 22:17:01 crc kubenswrapper[4931]: E1125 22:17:01.208024 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e\": container with ID starting with 70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e not found: ID does not exist" containerID="70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.208047 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e"} err="failed to get container status \"70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e\": rpc error: code = NotFound desc = could not find container \"70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e\": container with ID starting with 70b7aa38ed2a60bb6b8a9c5fbad3e567bc9a4569cade961b70b431a11409ad0e not found: ID does not exist" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.208061 4931 scope.go:117] "RemoveContainer" containerID="aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.208094 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kjrgw"] Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.222585 4931 scope.go:117] "RemoveContainer" containerID="f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.239982 4931 scope.go:117] "RemoveContainer" containerID="5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.257431 4931 scope.go:117] "RemoveContainer" containerID="aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9" Nov 25 22:17:01 crc kubenswrapper[4931]: E1125 22:17:01.257745 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9\": container with ID starting with aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9 not found: ID does not exist" containerID="aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.257770 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9"} err="failed to get container status \"aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9\": rpc error: code = NotFound desc = could not find container \"aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9\": container with ID starting with aa2e320953237a1f9d42097ac2f5855a5a21b19bd943eeb661a8d2408085efb9 not found: ID does not exist" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.257790 4931 scope.go:117] "RemoveContainer" containerID="f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7" Nov 25 22:17:01 crc kubenswrapper[4931]: E1125 22:17:01.258077 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7\": container with ID starting with f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7 not found: ID does not exist" containerID="f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.258095 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7"} err="failed to get container status \"f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7\": rpc error: code = NotFound desc = could not find container \"f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7\": container with ID starting with f0549560c1615ae1e75b8949d6194368fa89ee20671a85122ef16e49aaa687e7 not found: ID does not exist" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.258110 4931 scope.go:117] "RemoveContainer" containerID="5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98" Nov 25 22:17:01 crc kubenswrapper[4931]: E1125 22:17:01.258334 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98\": container with ID starting with 5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98 not found: ID does not exist" containerID="5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98" Nov 25 22:17:01 crc kubenswrapper[4931]: I1125 22:17:01.258357 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98"} err="failed to get container status \"5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98\": rpc error: code = NotFound desc = could not find container \"5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98\": container with ID starting with 5cf37c99b2da937e91d6a16a85c2b2ded52a5f73fc5d92d293bec1d8b5642f98 not found: ID does not exist" Nov 25 22:17:02 crc kubenswrapper[4931]: I1125 22:17:02.439792 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862" path="/var/lib/kubelet/pods/3bed4a4a-dc6f-4b3a-95d2-aa8ff5d8c862/volumes" Nov 25 22:17:02 crc kubenswrapper[4931]: I1125 22:17:02.441360 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a441c870-fc75-42d3-9c5a-2e432913db7f" path="/var/lib/kubelet/pods/a441c870-fc75-42d3-9c5a-2e432913db7f/volumes" Nov 25 22:17:04 crc kubenswrapper[4931]: I1125 22:17:04.032155 4931 scope.go:117] "RemoveContainer" containerID="85e8b871e841792add377f6341603951c2f6231fb5285d682bae046f7e4539e2" Nov 25 22:17:18 crc kubenswrapper[4931]: I1125 22:17:18.938453 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5_b355cbaf-1e51-49b8-a0a7-5591d6d62f25/util/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.096297 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5_b355cbaf-1e51-49b8-a0a7-5591d6d62f25/pull/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.105745 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5_b355cbaf-1e51-49b8-a0a7-5591d6d62f25/pull/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.134780 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5_b355cbaf-1e51-49b8-a0a7-5591d6d62f25/util/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.294401 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5_b355cbaf-1e51-49b8-a0a7-5591d6d62f25/util/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.323121 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5_b355cbaf-1e51-49b8-a0a7-5591d6d62f25/pull/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.343959 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fpgzp5_b355cbaf-1e51-49b8-a0a7-5591d6d62f25/extract/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.455858 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq_bdfe8e93-8490-4958-aab7-0c2b062b22fa/util/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.628258 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq_bdfe8e93-8490-4958-aab7-0c2b062b22fa/pull/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.628849 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq_bdfe8e93-8490-4958-aab7-0c2b062b22fa/util/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.629965 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq_bdfe8e93-8490-4958-aab7-0c2b062b22fa/pull/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.813048 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq_bdfe8e93-8490-4958-aab7-0c2b062b22fa/pull/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.826691 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq_bdfe8e93-8490-4958-aab7-0c2b062b22fa/util/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.831152 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5c7190afbb317be1c8302d452aac5b107c23692e596cfc5d32fe58ad23c99rq_bdfe8e93-8490-4958-aab7-0c2b062b22fa/extract/0.log" Nov 25 22:17:19 crc kubenswrapper[4931]: I1125 22:17:19.966054 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg_a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c/util/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.129976 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg_a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c/pull/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.144749 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg_a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c/util/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.159954 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg_a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c/pull/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.295485 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg_a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c/util/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.315218 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg_a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c/pull/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.327227 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5djvmsg_a45f74e7-7adf-47d7-bdf1-7e9ad4a48a3c/extract/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.474440 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2_4237875e-d36d-4c45-9102-2f1242407c46/util/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.723679 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2_4237875e-d36d-4c45-9102-2f1242407c46/pull/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.741838 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2_4237875e-d36d-4c45-9102-2f1242407c46/util/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.751046 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2_4237875e-d36d-4c45-9102-2f1242407c46/pull/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.930125 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2_4237875e-d36d-4c45-9102-2f1242407c46/util/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.936204 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2_4237875e-d36d-4c45-9102-2f1242407c46/pull/0.log" Nov 25 22:17:20 crc kubenswrapper[4931]: I1125 22:17:20.957611 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bk69z2_4237875e-d36d-4c45-9102-2f1242407c46/extract/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.086447 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6_b979640a-23b5-48df-96be-82bc4630d172/util/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.258181 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6_b979640a-23b5-48df-96be-82bc4630d172/pull/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.298785 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6_b979640a-23b5-48df-96be-82bc4630d172/pull/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.348655 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6_b979640a-23b5-48df-96be-82bc4630d172/util/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.525443 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6_b979640a-23b5-48df-96be-82bc4630d172/pull/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.580937 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6_b979640a-23b5-48df-96be-82bc4630d172/util/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.585548 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907pqb6_b979640a-23b5-48df-96be-82bc4630d172/extract/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.599890 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8_a748bed2-f75d-4cfb-b20d-3914cb4094ee/util/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.716704 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8_a748bed2-f75d-4cfb-b20d-3914cb4094ee/pull/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.731985 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8_a748bed2-f75d-4cfb-b20d-3914cb4094ee/util/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.748243 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8_a748bed2-f75d-4cfb-b20d-3914cb4094ee/pull/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.919493 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8_a748bed2-f75d-4cfb-b20d-3914cb4094ee/util/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.938115 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8_a748bed2-f75d-4cfb-b20d-3914cb4094ee/extract/0.log" Nov 25 22:17:21 crc kubenswrapper[4931]: I1125 22:17:21.943402 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c667s8_a748bed2-f75d-4cfb-b20d-3914cb4094ee/pull/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.030859 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24_3039d342-7a69-4ad3-bd7e-e21a65264f61/util/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.154857 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24_3039d342-7a69-4ad3-bd7e-e21a65264f61/util/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.174298 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24_3039d342-7a69-4ad3-bd7e-e21a65264f61/pull/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.185191 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24_3039d342-7a69-4ad3-bd7e-e21a65264f61/pull/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.315169 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24_3039d342-7a69-4ad3-bd7e-e21a65264f61/util/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.335258 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24_3039d342-7a69-4ad3-bd7e-e21a65264f61/pull/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.351511 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3qhl24_3039d342-7a69-4ad3-bd7e-e21a65264f61/extract/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.369111 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-6fb5bb54f-chfmg_0700c2d1-0a31-4d46-9fe7-4bdb809d1cef/manager/3.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.491191 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-6fb5bb54f-chfmg_0700c2d1-0a31-4d46-9fe7-4bdb809d1cef/manager/2.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.515768 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-svnj9_559c2f5b-32bc-4a0e-95ef-d7f4ad7a902a/registry-server/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.555014 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6bfcf864bd-cfmpv_4e0590b1-c897-4382-bc93-4a1b0653a526/manager/3.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.657355 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6bfcf864bd-cfmpv_4e0590b1-c897-4382-bc93-4a1b0653a526/manager/2.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.716142 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-mhxw4_502b379e-39c4-4a24-a639-7aa82885c7fd/registry-server/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.760577 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6545c6c9d6-7rkzq_f4575978-a78e-42b2-a2dc-1052e5777b04/kube-rbac-proxy/0.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.854903 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6545c6c9d6-7rkzq_f4575978-a78e-42b2-a2dc-1052e5777b04/manager/3.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.876385 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6545c6c9d6-7rkzq_f4575978-a78e-42b2-a2dc-1052e5777b04/manager/2.log" Nov 25 22:17:22 crc kubenswrapper[4931]: I1125 22:17:22.936375 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-kv5fj_3f1ba234-ef51-46fd-81ba-16db04e0c689/registry-server/0.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.055373 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b59697cf5-jthhc_044c2639-0ceb-4be3-8326-4fdc552aefb6/manager/3.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.061571 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b59697cf5-jthhc_044c2639-0ceb-4be3-8326-4fdc552aefb6/manager/2.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.100461 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-8gl8x_d83baa8b-fccc-4dd8-9fd9-f7fe5850e5ca/registry-server/0.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.228629 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-7b69997d9b-sz4qv_952695c7-9eff-4e33-8777-6a5ea50a2b94/manager/3.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.248652 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-7b69997d9b-sz4qv_952695c7-9eff-4e33-8777-6a5ea50a2b94/manager/2.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.341134 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-znxs7_413e27a3-fd17-4743-9ac4-6767858ea041/registry-server/0.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.404603 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-zsd98_7b1bb711-ce82-4ea4-9667-ff4816b162f4/operator/3.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.423827 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-zsd98_7b1bb711-ce82-4ea4-9667-ff4816b162f4/operator/2.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.484358 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-t6lmz_ad908caf-8931-44bb-a5b3-a59952dfc01a/registry-server/0.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.587347 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7cb66c858-rgrsz_8f67c56d-e216-435c-8384-91d28a4d864e/manager/2.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.608618 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7cb66c858-rgrsz_8f67c56d-e216-435c-8384-91d28a4d864e/manager/3.log" Nov 25 22:17:23 crc kubenswrapper[4931]: I1125 22:17:23.691673 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-4v74n_6162ec0d-ae9b-4334-a7d0-bc693ee687d2/registry-server/0.log" Nov 25 22:17:39 crc kubenswrapper[4931]: I1125 22:17:39.549276 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-rsbdr_45811a4b-98c3-4f54-b772-f1adc68d88c7/control-plane-machine-set-operator/0.log" Nov 25 22:17:39 crc kubenswrapper[4931]: I1125 22:17:39.670475 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5dlk4_842021b8-6f3c-4eab-bcba-0b27f6736e9e/kube-rbac-proxy/0.log" Nov 25 22:17:39 crc kubenswrapper[4931]: I1125 22:17:39.701950 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5dlk4_842021b8-6f3c-4eab-bcba-0b27f6736e9e/machine-api-operator/0.log" Nov 25 22:17:54 crc kubenswrapper[4931]: I1125 22:17:54.063103 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-wntmn"] Nov 25 22:17:54 crc kubenswrapper[4931]: I1125 22:17:54.069323 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-3765-account-create-update-bvwnb"] Nov 25 22:17:54 crc kubenswrapper[4931]: I1125 22:17:54.076481 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-3765-account-create-update-bvwnb"] Nov 25 22:17:54 crc kubenswrapper[4931]: I1125 22:17:54.082359 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-wntmn"] Nov 25 22:17:54 crc kubenswrapper[4931]: I1125 22:17:54.435177 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ef44b19-30a3-4d00-ae70-78dc171bc820" path="/var/lib/kubelet/pods/4ef44b19-30a3-4d00-ae70-78dc171bc820/volumes" Nov 25 22:17:54 crc kubenswrapper[4931]: I1125 22:17:54.436712 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6933bb9-7603-4a25-bc47-685a2e75ce2b" path="/var/lib/kubelet/pods/a6933bb9-7603-4a25-bc47-685a2e75ce2b/volumes" Nov 25 22:17:56 crc kubenswrapper[4931]: I1125 22:17:56.929701 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-xhr6z_c77cee34-eac7-4391-8fc0-f492acaa9004/kube-rbac-proxy/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.078099 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-xhr6z_c77cee34-eac7-4391-8fc0-f492acaa9004/controller/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.122722 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-frr-files/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.293843 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-frr-files/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.293893 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-reloader/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.318450 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-metrics/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.334268 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-reloader/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.472196 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-frr-files/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.474596 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-metrics/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.480083 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-reloader/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.517947 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-metrics/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.699859 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-metrics/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.708018 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-reloader/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.709667 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/cp-frr-files/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.727296 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/controller/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.880157 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/frr-metrics/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.899398 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/kube-rbac-proxy/0.log" Nov 25 22:17:57 crc kubenswrapper[4931]: I1125 22:17:57.950097 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/kube-rbac-proxy-frr/0.log" Nov 25 22:17:58 crc kubenswrapper[4931]: I1125 22:17:58.100267 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/reloader/0.log" Nov 25 22:17:58 crc kubenswrapper[4931]: I1125 22:17:58.220640 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wcxdv_b3601564-28ae-4877-8b80-33b5027bfcf5/frr/0.log" Nov 25 22:17:58 crc kubenswrapper[4931]: I1125 22:17:58.230469 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-2wqcf_fd84f23e-59df-4222-b73d-e291f7a18301/frr-k8s-webhook-server/0.log" Nov 25 22:17:58 crc kubenswrapper[4931]: I1125 22:17:58.317135 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5bfcf68b5b-79jjz_60182b4c-52b4-47a9-8c5a-ab0274136ef1/manager/3.log" Nov 25 22:17:58 crc kubenswrapper[4931]: I1125 22:17:58.408206 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5bfcf68b5b-79jjz_60182b4c-52b4-47a9-8c5a-ab0274136ef1/manager/2.log" Nov 25 22:17:58 crc kubenswrapper[4931]: I1125 22:17:58.465215 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6dcc996cd8-f5vhn_ba8977d1-518a-46c8-b174-073b37d6ebb6/webhook-server/0.log" Nov 25 22:17:58 crc kubenswrapper[4931]: I1125 22:17:58.573347 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qsmp8_15968cf8-d97e-4ba4-95c5-a3d82eb43e9c/kube-rbac-proxy/0.log" Nov 25 22:17:58 crc kubenswrapper[4931]: I1125 22:17:58.719697 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qsmp8_15968cf8-d97e-4ba4-95c5-a3d82eb43e9c/speaker/0.log" Nov 25 22:18:04 crc kubenswrapper[4931]: I1125 22:18:04.133312 4931 scope.go:117] "RemoveContainer" containerID="1ddc3349c2ae687066755465d87ec9e6bccadcfbc00f4358dd9f937123d2a87f" Nov 25 22:18:04 crc kubenswrapper[4931]: I1125 22:18:04.174702 4931 scope.go:117] "RemoveContainer" containerID="104ceb20bf6a92e6ce2ee2924f41be0756130e1eb6de72fbb776f7c2989ac896" Nov 25 22:18:04 crc kubenswrapper[4931]: I1125 22:18:04.203828 4931 scope.go:117] "RemoveContainer" containerID="3b241f4f0d61953fd8db9f03162f41c15f4640b20ed666996d54bbdd4f451966" Nov 25 22:18:08 crc kubenswrapper[4931]: I1125 22:18:08.844076 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:18:08 crc kubenswrapper[4931]: I1125 22:18:08.844701 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:18:09 crc kubenswrapper[4931]: I1125 22:18:09.046450 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-ngsxr"] Nov 25 22:18:09 crc kubenswrapper[4931]: I1125 22:18:09.053489 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-ngsxr"] Nov 25 22:18:10 crc kubenswrapper[4931]: I1125 22:18:10.432961 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="575383b0-d5a9-41a6-8799-f7cd7e3e637f" path="/var/lib/kubelet/pods/575383b0-d5a9-41a6-8799-f7cd7e3e637f/volumes" Nov 25 22:18:13 crc kubenswrapper[4931]: I1125 22:18:13.926102 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-6a8c-account-create-update-t2mvt_aa62d794-8196-47d9-bf6b-515ea4d1a35b/mariadb-account-create-update/0.log" Nov 25 22:18:13 crc kubenswrapper[4931]: I1125 22:18:13.990061 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-create-8j7l7_a3836796-a01c-4f07-a2e9-7dfc2b7054e1/mariadb-database-create/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.109188 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-5h92t_5d7d25ea-0573-484d-bd94-e9f1527edf1f/glance-db-sync/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.182386 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_0987f7d6-37f0-48ea-8873-13cfa65fde9c/glance-api/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.272231 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_0987f7d6-37f0-48ea-8873-13cfa65fde9c/glance-httpd/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.315622 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_0987f7d6-37f0-48ea-8873-13cfa65fde9c/glance-log/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.356510 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-1_e68e5a87-bc0c-4683-a9c2-aac189ca3c6d/glance-api/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.484619 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-1_e68e5a87-bc0c-4683-a9c2-aac189ca3c6d/glance-log/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.489735 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-1_e68e5a87-bc0c-4683-a9c2-aac189ca3c6d/glance-httpd/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.533544 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-2_28023b78-ba9e-4dbe-9033-59df28b3437c/glance-api/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.671226 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-2_28023b78-ba9e-4dbe-9033-59df28b3437c/glance-log/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.680086 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-2_28023b78-ba9e-4dbe-9033-59df28b3437c/glance-httpd/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.734266 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_b7c2ca26-5323-4f31-ba89-a529dd5ddde0/glance-api/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.832380 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_b7c2ca26-5323-4f31-ba89-a529dd5ddde0/glance-httpd/0.log" Nov 25 22:18:14 crc kubenswrapper[4931]: I1125 22:18:14.873307 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_b7c2ca26-5323-4f31-ba89-a529dd5ddde0/glance-log/0.log" Nov 25 22:18:15 crc kubenswrapper[4931]: I1125 22:18:15.055559 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-bootstrap-4s625_200fa711-1e82-4671-8281-555f5139ee24/keystone-bootstrap/0.log" Nov 25 22:18:15 crc kubenswrapper[4931]: I1125 22:18:15.095591 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-8d4cbfbc4-b5kz7_6e715806-4ba6-436a-b623-1403898fd27a/keystone-api/0.log" Nov 25 22:18:15 crc kubenswrapper[4931]: I1125 22:18:15.255153 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb/mysql-bootstrap/0.log" Nov 25 22:18:15 crc kubenswrapper[4931]: I1125 22:18:15.478315 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb/galera/0.log" Nov 25 22:18:15 crc kubenswrapper[4931]: I1125 22:18:15.526433 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_56fd8864-ec73-42e3-8dd9-0d9f5d29d7eb/mysql-bootstrap/0.log" Nov 25 22:18:15 crc kubenswrapper[4931]: I1125 22:18:15.688574 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_fe87b340-3470-4f3e-a69a-e041cdcd24c2/mysql-bootstrap/0.log" Nov 25 22:18:15 crc kubenswrapper[4931]: I1125 22:18:15.893584 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_fe87b340-3470-4f3e-a69a-e041cdcd24c2/galera/0.log" Nov 25 22:18:15 crc kubenswrapper[4931]: I1125 22:18:15.894815 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_fe87b340-3470-4f3e-a69a-e041cdcd24c2/mysql-bootstrap/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.025954 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4s625"] Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.052009 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4s625"] Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.071311 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a/mysql-bootstrap/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.220688 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_27505da7-a979-42ac-9572-b8a2f6a075de/memcached/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.305123 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a/mysql-bootstrap/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.318963 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_c7a1ed06-d8b3-42ef-9fb0-795bbc337b3a/galera/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.395829 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_e5f137dc-0889-45c9-bfb7-f99227ae93df/openstackclient/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.430410 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="200fa711-1e82-4671-8281-555f5139ee24" path="/var/lib/kubelet/pods/200fa711-1e82-4671-8281-555f5139ee24/volumes" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.511546 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_4b3d80c3-c790-47e0-ae6c-5c639da2ef84/setup-container/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.614716 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_4b3d80c3-c790-47e0-ae6c-5c639da2ef84/rabbitmq/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.633493 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_4b3d80c3-c790-47e0-ae6c-5c639da2ef84/setup-container/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.748763 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-4j5sd_fb416fb6-2f7e-405f-aba4-588fa1929bbd/proxy-httpd/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.752409 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-4j5sd_fb416fb6-2f7e-405f-aba4-588fa1929bbd/proxy-server/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.808221 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-w2kg9_796108c0-b88f-4a48-8596-15155c91b126/swift-ring-rebalance/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.927008 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/account-auditor/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.962152 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/account-server/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.963792 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/account-reaper/0.log" Nov 25 22:18:16 crc kubenswrapper[4931]: I1125 22:18:16.989737 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/account-replicator/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.321292 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/container-auditor/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.406388 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/container-replicator/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.477263 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/container-server/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.477722 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/container-updater/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.527804 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/object-auditor/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.530743 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/object-expirer/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.603657 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/object-replicator/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.639509 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/object-updater/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.663433 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/object-server/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.803758 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/rsync/0.log" Nov 25 22:18:17 crc kubenswrapper[4931]: I1125 22:18:17.810819 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_2f56d793-2122-4466-8dda-38687d951a74/swift-recon-cron/0.log" Nov 25 22:18:32 crc kubenswrapper[4931]: I1125 22:18:32.596083 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fm7z9_acbf2582-f183-485f-8947-7b8b5cd4dd75/extract-utilities/0.log" Nov 25 22:18:32 crc kubenswrapper[4931]: I1125 22:18:32.784636 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fm7z9_acbf2582-f183-485f-8947-7b8b5cd4dd75/extract-utilities/0.log" Nov 25 22:18:32 crc kubenswrapper[4931]: I1125 22:18:32.785321 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fm7z9_acbf2582-f183-485f-8947-7b8b5cd4dd75/extract-content/0.log" Nov 25 22:18:32 crc kubenswrapper[4931]: I1125 22:18:32.812074 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fm7z9_acbf2582-f183-485f-8947-7b8b5cd4dd75/extract-content/0.log" Nov 25 22:18:32 crc kubenswrapper[4931]: I1125 22:18:32.976281 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fm7z9_acbf2582-f183-485f-8947-7b8b5cd4dd75/extract-content/0.log" Nov 25 22:18:32 crc kubenswrapper[4931]: I1125 22:18:32.979377 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fm7z9_acbf2582-f183-485f-8947-7b8b5cd4dd75/extract-utilities/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.183791 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hhsmv_372612f8-655b-4bbf-a36c-52b59ce3dda0/extract-utilities/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.360649 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fm7z9_acbf2582-f183-485f-8947-7b8b5cd4dd75/registry-server/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.369926 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hhsmv_372612f8-655b-4bbf-a36c-52b59ce3dda0/extract-utilities/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.395911 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hhsmv_372612f8-655b-4bbf-a36c-52b59ce3dda0/extract-content/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.399999 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hhsmv_372612f8-655b-4bbf-a36c-52b59ce3dda0/extract-content/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.530204 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hhsmv_372612f8-655b-4bbf-a36c-52b59ce3dda0/extract-utilities/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.540830 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hhsmv_372612f8-655b-4bbf-a36c-52b59ce3dda0/extract-content/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.731783 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94/util/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.940202 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hhsmv_372612f8-655b-4bbf-a36c-52b59ce3dda0/registry-server/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.947181 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94/util/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.947620 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94/pull/0.log" Nov 25 22:18:33 crc kubenswrapper[4931]: I1125 22:18:33.976094 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94/pull/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.101934 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94/extract/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.114043 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94/util/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.152861 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67kq8j_79cfa2ed-ad95-45eb-a55d-3f1fb6f08e94/pull/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.256270 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hffz8_d0b65dd4-c1b3-4175-9951-3952c2fc4186/marketplace-operator/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.263024 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hffz8_d0b65dd4-c1b3-4175-9951-3952c2fc4186/marketplace-operator/1.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.310475 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5g982_a47b9365-6b25-4ce9-8f07-96afc03a8634/extract-utilities/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.509170 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5g982_a47b9365-6b25-4ce9-8f07-96afc03a8634/extract-content/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.519713 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5g982_a47b9365-6b25-4ce9-8f07-96afc03a8634/extract-content/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.553114 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5g982_a47b9365-6b25-4ce9-8f07-96afc03a8634/extract-utilities/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.683712 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5g982_a47b9365-6b25-4ce9-8f07-96afc03a8634/extract-content/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.767234 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5g982_a47b9365-6b25-4ce9-8f07-96afc03a8634/extract-utilities/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.808443 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5g982_a47b9365-6b25-4ce9-8f07-96afc03a8634/registry-server/0.log" Nov 25 22:18:34 crc kubenswrapper[4931]: I1125 22:18:34.879494 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhh6r_ddf049ad-4aee-4c0a-ab15-6c4ef560a755/extract-utilities/0.log" Nov 25 22:18:35 crc kubenswrapper[4931]: I1125 22:18:35.003986 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhh6r_ddf049ad-4aee-4c0a-ab15-6c4ef560a755/extract-utilities/0.log" Nov 25 22:18:35 crc kubenswrapper[4931]: I1125 22:18:35.028805 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhh6r_ddf049ad-4aee-4c0a-ab15-6c4ef560a755/extract-content/0.log" Nov 25 22:18:35 crc kubenswrapper[4931]: I1125 22:18:35.041049 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhh6r_ddf049ad-4aee-4c0a-ab15-6c4ef560a755/extract-content/0.log" Nov 25 22:18:35 crc kubenswrapper[4931]: I1125 22:18:35.191904 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhh6r_ddf049ad-4aee-4c0a-ab15-6c4ef560a755/extract-utilities/0.log" Nov 25 22:18:35 crc kubenswrapper[4931]: I1125 22:18:35.195590 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhh6r_ddf049ad-4aee-4c0a-ab15-6c4ef560a755/extract-content/0.log" Nov 25 22:18:35 crc kubenswrapper[4931]: I1125 22:18:35.547071 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jhh6r_ddf049ad-4aee-4c0a-ab15-6c4ef560a755/registry-server/0.log" Nov 25 22:18:38 crc kubenswrapper[4931]: I1125 22:18:38.843852 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:18:38 crc kubenswrapper[4931]: I1125 22:18:38.844179 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:19:04 crc kubenswrapper[4931]: I1125 22:19:04.281341 4931 scope.go:117] "RemoveContainer" containerID="97984b2543d5eebb0b54e5e7de0c74158a096a1064e7203da78e19ab861cfe8e" Nov 25 22:19:04 crc kubenswrapper[4931]: I1125 22:19:04.345147 4931 scope.go:117] "RemoveContainer" containerID="1685fe75f11d8354658228b93dcf6f13b6b974e4bc838416f7c599769371439e" Nov 25 22:19:08 crc kubenswrapper[4931]: I1125 22:19:08.843453 4931 patch_prober.go:28] interesting pod/machine-config-daemon-djjhg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 22:19:08 crc kubenswrapper[4931]: I1125 22:19:08.843849 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 22:19:08 crc kubenswrapper[4931]: I1125 22:19:08.843941 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" Nov 25 22:19:08 crc kubenswrapper[4931]: I1125 22:19:08.845119 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429"} pod="openshift-machine-config-operator/machine-config-daemon-djjhg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 22:19:08 crc kubenswrapper[4931]: I1125 22:19:08.845209 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" containerName="machine-config-daemon" containerID="cri-o://0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" gracePeriod=600 Nov 25 22:19:09 crc kubenswrapper[4931]: E1125 22:19:09.013475 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:19:09 crc kubenswrapper[4931]: I1125 22:19:09.283323 4931 generic.go:334] "Generic (PLEG): container finished" podID="e22448de-5122-43be-853a-081347c13cf2" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" exitCode=0 Nov 25 22:19:09 crc kubenswrapper[4931]: I1125 22:19:09.283385 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" event={"ID":"e22448de-5122-43be-853a-081347c13cf2","Type":"ContainerDied","Data":"0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429"} Nov 25 22:19:09 crc kubenswrapper[4931]: I1125 22:19:09.283740 4931 scope.go:117] "RemoveContainer" containerID="90db9e7682e94b8d54bbdd5d154989aeb94726c14a1c13fff2838dc83db574a2" Nov 25 22:19:09 crc kubenswrapper[4931]: I1125 22:19:09.284662 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:19:09 crc kubenswrapper[4931]: E1125 22:19:09.285144 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:19:20 crc kubenswrapper[4931]: I1125 22:19:20.422708 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:19:20 crc kubenswrapper[4931]: E1125 22:19:20.423862 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:19:34 crc kubenswrapper[4931]: I1125 22:19:34.422765 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:19:34 crc kubenswrapper[4931]: E1125 22:19:34.424056 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:19:39 crc kubenswrapper[4931]: I1125 22:19:39.561852 4931 generic.go:334] "Generic (PLEG): container finished" podID="aea87457-cffd-4a2d-a368-4c5efb5941ac" containerID="87a8dc8ef076c36dbd3dd63f8f5d2fbc9255f32469bbb50df6a85c78ae825d70" exitCode=0 Nov 25 22:19:39 crc kubenswrapper[4931]: I1125 22:19:39.561949 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" event={"ID":"aea87457-cffd-4a2d-a368-4c5efb5941ac","Type":"ContainerDied","Data":"87a8dc8ef076c36dbd3dd63f8f5d2fbc9255f32469bbb50df6a85c78ae825d70"} Nov 25 22:19:39 crc kubenswrapper[4931]: I1125 22:19:39.563244 4931 scope.go:117] "RemoveContainer" containerID="87a8dc8ef076c36dbd3dd63f8f5d2fbc9255f32469bbb50df6a85c78ae825d70" Nov 25 22:19:40 crc kubenswrapper[4931]: I1125 22:19:40.369071 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksnv8_must-gather-hbhpw_aea87457-cffd-4a2d-a368-4c5efb5941ac/gather/0.log" Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.317374 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ksnv8/must-gather-hbhpw"] Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.318544 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" podUID="aea87457-cffd-4a2d-a368-4c5efb5941ac" containerName="copy" containerID="cri-o://8ff403f34d8bc6d8bca03a3041c257199bf830ac1d3815cc785d53c0d8d92fdd" gracePeriod=2 Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.329817 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ksnv8/must-gather-hbhpw"] Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.423076 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:19:47 crc kubenswrapper[4931]: E1125 22:19:47.423410 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.638395 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksnv8_must-gather-hbhpw_aea87457-cffd-4a2d-a368-4c5efb5941ac/copy/0.log" Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.639771 4931 generic.go:334] "Generic (PLEG): container finished" podID="aea87457-cffd-4a2d-a368-4c5efb5941ac" containerID="8ff403f34d8bc6d8bca03a3041c257199bf830ac1d3815cc785d53c0d8d92fdd" exitCode=143 Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.689757 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksnv8_must-gather-hbhpw_aea87457-cffd-4a2d-a368-4c5efb5941ac/copy/0.log" Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.690215 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.717781 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xwx2\" (UniqueName: \"kubernetes.io/projected/aea87457-cffd-4a2d-a368-4c5efb5941ac-kube-api-access-9xwx2\") pod \"aea87457-cffd-4a2d-a368-4c5efb5941ac\" (UID: \"aea87457-cffd-4a2d-a368-4c5efb5941ac\") " Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.717913 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aea87457-cffd-4a2d-a368-4c5efb5941ac-must-gather-output\") pod \"aea87457-cffd-4a2d-a368-4c5efb5941ac\" (UID: \"aea87457-cffd-4a2d-a368-4c5efb5941ac\") " Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.727318 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aea87457-cffd-4a2d-a368-4c5efb5941ac-kube-api-access-9xwx2" (OuterVolumeSpecName: "kube-api-access-9xwx2") pod "aea87457-cffd-4a2d-a368-4c5efb5941ac" (UID: "aea87457-cffd-4a2d-a368-4c5efb5941ac"). InnerVolumeSpecName "kube-api-access-9xwx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.797380 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aea87457-cffd-4a2d-a368-4c5efb5941ac-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "aea87457-cffd-4a2d-a368-4c5efb5941ac" (UID: "aea87457-cffd-4a2d-a368-4c5efb5941ac"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.821177 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xwx2\" (UniqueName: \"kubernetes.io/projected/aea87457-cffd-4a2d-a368-4c5efb5941ac-kube-api-access-9xwx2\") on node \"crc\" DevicePath \"\"" Nov 25 22:19:47 crc kubenswrapper[4931]: I1125 22:19:47.821229 4931 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aea87457-cffd-4a2d-a368-4c5efb5941ac-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 22:19:48 crc kubenswrapper[4931]: I1125 22:19:48.433864 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aea87457-cffd-4a2d-a368-4c5efb5941ac" path="/var/lib/kubelet/pods/aea87457-cffd-4a2d-a368-4c5efb5941ac/volumes" Nov 25 22:19:48 crc kubenswrapper[4931]: I1125 22:19:48.657194 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksnv8_must-gather-hbhpw_aea87457-cffd-4a2d-a368-4c5efb5941ac/copy/0.log" Nov 25 22:19:48 crc kubenswrapper[4931]: I1125 22:19:48.657685 4931 scope.go:117] "RemoveContainer" containerID="8ff403f34d8bc6d8bca03a3041c257199bf830ac1d3815cc785d53c0d8d92fdd" Nov 25 22:19:48 crc kubenswrapper[4931]: I1125 22:19:48.657755 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksnv8/must-gather-hbhpw" Nov 25 22:19:48 crc kubenswrapper[4931]: I1125 22:19:48.677648 4931 scope.go:117] "RemoveContainer" containerID="87a8dc8ef076c36dbd3dd63f8f5d2fbc9255f32469bbb50df6a85c78ae825d70" Nov 25 22:19:50 crc kubenswrapper[4931]: E1125 22:19:50.924038 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Nov 25 22:19:59 crc kubenswrapper[4931]: I1125 22:19:59.423681 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:19:59 crc kubenswrapper[4931]: E1125 22:19:59.424829 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:20:13 crc kubenswrapper[4931]: I1125 22:20:13.423366 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:20:13 crc kubenswrapper[4931]: E1125 22:20:13.424104 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:20:24 crc kubenswrapper[4931]: I1125 22:20:24.426256 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:20:24 crc kubenswrapper[4931]: E1125 22:20:24.427243 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:20:36 crc kubenswrapper[4931]: I1125 22:20:36.423540 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:20:36 crc kubenswrapper[4931]: E1125 22:20:36.424857 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:20:51 crc kubenswrapper[4931]: I1125 22:20:51.423057 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:20:51 crc kubenswrapper[4931]: E1125 22:20:51.424000 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:21:05 crc kubenswrapper[4931]: I1125 22:21:05.423330 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:21:05 crc kubenswrapper[4931]: E1125 22:21:05.424269 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:21:18 crc kubenswrapper[4931]: I1125 22:21:18.424312 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:21:18 crc kubenswrapper[4931]: E1125 22:21:18.425371 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:21:28 crc kubenswrapper[4931]: I1125 22:21:28.039193 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-8j7l7"] Nov 25 22:21:28 crc kubenswrapper[4931]: I1125 22:21:28.048026 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-8j7l7"] Nov 25 22:21:28 crc kubenswrapper[4931]: I1125 22:21:28.434458 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3836796-a01c-4f07-a2e9-7dfc2b7054e1" path="/var/lib/kubelet/pods/a3836796-a01c-4f07-a2e9-7dfc2b7054e1/volumes" Nov 25 22:21:29 crc kubenswrapper[4931]: I1125 22:21:29.034690 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt"] Nov 25 22:21:29 crc kubenswrapper[4931]: I1125 22:21:29.049870 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-6a8c-account-create-update-t2mvt"] Nov 25 22:21:30 crc kubenswrapper[4931]: I1125 22:21:30.439313 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa62d794-8196-47d9-bf6b-515ea4d1a35b" path="/var/lib/kubelet/pods/aa62d794-8196-47d9-bf6b-515ea4d1a35b/volumes" Nov 25 22:21:33 crc kubenswrapper[4931]: I1125 22:21:33.422913 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:21:33 crc kubenswrapper[4931]: E1125 22:21:33.423609 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:21:37 crc kubenswrapper[4931]: I1125 22:21:37.047280 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5h92t"] Nov 25 22:21:37 crc kubenswrapper[4931]: I1125 22:21:37.059749 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5h92t"] Nov 25 22:21:38 crc kubenswrapper[4931]: I1125 22:21:38.443854 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d7d25ea-0573-484d-bd94-e9f1527edf1f" path="/var/lib/kubelet/pods/5d7d25ea-0573-484d-bd94-e9f1527edf1f/volumes" Nov 25 22:21:46 crc kubenswrapper[4931]: I1125 22:21:46.423292 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:21:46 crc kubenswrapper[4931]: E1125 22:21:46.424445 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:21:59 crc kubenswrapper[4931]: I1125 22:21:59.422520 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:21:59 crc kubenswrapper[4931]: E1125 22:21:59.423513 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:22:04 crc kubenswrapper[4931]: I1125 22:22:04.524903 4931 scope.go:117] "RemoveContainer" containerID="bbe4cbf4f8bd07253e3590f255203ff3eb6ca5585c0a97e5c353f16a7abc5ea4" Nov 25 22:22:04 crc kubenswrapper[4931]: I1125 22:22:04.563929 4931 scope.go:117] "RemoveContainer" containerID="bb98a279a2d470505b2d878082f2849be02eb50a03d8bfeaaa8637c0540fa30c" Nov 25 22:22:04 crc kubenswrapper[4931]: I1125 22:22:04.639202 4931 scope.go:117] "RemoveContainer" containerID="902741845ac8789170899ec25fcba35ce5e4095b888512b806c7823cbca65d73" Nov 25 22:22:14 crc kubenswrapper[4931]: I1125 22:22:14.423137 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:22:14 crc kubenswrapper[4931]: E1125 22:22:14.424342 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:22:29 crc kubenswrapper[4931]: I1125 22:22:29.423335 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:22:29 crc kubenswrapper[4931]: E1125 22:22:29.424320 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:22:40 crc kubenswrapper[4931]: I1125 22:22:40.427821 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:22:40 crc kubenswrapper[4931]: E1125 22:22:40.428594 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:22:54 crc kubenswrapper[4931]: I1125 22:22:54.427412 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:22:54 crc kubenswrapper[4931]: E1125 22:22:54.428389 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:23:09 crc kubenswrapper[4931]: I1125 22:23:09.423170 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:23:09 crc kubenswrapper[4931]: E1125 22:23:09.424331 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:23:20 crc kubenswrapper[4931]: I1125 22:23:20.423011 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:23:20 crc kubenswrapper[4931]: E1125 22:23:20.423950 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:23:33 crc kubenswrapper[4931]: I1125 22:23:33.422308 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:23:33 crc kubenswrapper[4931]: E1125 22:23:33.423200 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" Nov 25 22:23:48 crc kubenswrapper[4931]: I1125 22:23:48.422852 4931 scope.go:117] "RemoveContainer" containerID="0ee83c82bc9943d9fc833a95f2245035f28a592eedc3fd7a4a9855d32cd06429" Nov 25 22:23:48 crc kubenswrapper[4931]: E1125 22:23:48.424024 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djjhg_openshift-machine-config-operator(e22448de-5122-43be-853a-081347c13cf2)\"" pod="openshift-machine-config-operator/machine-config-daemon-djjhg" podUID="e22448de-5122-43be-853a-081347c13cf2" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111426203024440 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111426203017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111421770016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111421770015454 5ustar corecore